2 research outputs found
Contrastive Deep Encoding Enables Uncertainty-aware Machine-learning-assisted Histopathology
Deep neural network models can learn clinically relevant features from
millions of histopathology images. However generating high-quality annotations
to train such models for each hospital, each cancer type, and each diagnostic
task is prohibitively laborious. On the other hand, terabytes of training data
-- while lacking reliable annotations -- are readily available in the public
domain in some cases. In this work, we explore how these large datasets can be
consciously utilized to pre-train deep networks to encode informative
representations. We then fine-tune our pre-trained models on a fraction of
annotated training data to perform specific downstream tasks. We show that our
approach can reach the state-of-the-art (SOTA) for patch-level classification
with only 1-10% randomly selected annotations compared to other SOTA
approaches. Moreover, we propose an uncertainty-aware loss function, to
quantify the model confidence during inference. Quantified uncertainty helps
experts select the best instances to label for further training. Our
uncertainty-aware labeling reaches the SOTA with significantly fewer
annotations compared to random labeling. Last, we demonstrate how our
pre-trained encoders can surpass current SOTA for whole-slide image
classification with weak supervision. Our work lays the foundation for data and
task-agnostic pre-trained deep networks with quantified uncertainty.Comment: 18 pages, 8 figure
Hierarchical Text-to-Vision Self Supervised Alignment for Improved Histopathology Representation Learning
Self-supervised representation learning has been highly promising for
histopathology image analysis with numerous approaches leveraging their
patient-slide-patch hierarchy to learn better representations. In this paper,
we explore how the combination of domain specific natural language information
with such hierarchical visual representations can benefit rich representation
learning for medical image tasks. Building on automated language description
generation for features visible in histopathology images, we present a novel
language-tied self-supervised learning framework, Hierarchical Language-tied
Self-Supervision (HLSS) for histopathology images. We explore contrastive
objectives and granular language description based text alignment at multiple
hierarchies to inject language modality information into the visual
representations. Our resulting model achieves state-of-the-art performance on
two medical imaging benchmarks, OpenSRH and TCGA datasets. Our framework also
provides better interpretability with our language aligned representation
space. Code is available at https://github.com/Hasindri/HLSS.Comment: 13 pages and 5 figure