6,366 research outputs found
G2C: A Generator-to-Classifier Framework Integrating Multi-Stained Visual Cues for Pathological Glomerulus Classification
Pathological glomerulus classification plays a key role in the diagnosis of
nephropathy. As the difference between different subcategories is subtle,
doctors often refer to slides from different staining methods to make
decisions. However, creating correspondence across various stains is
labor-intensive, bringing major difficulties in collecting data and training a
vision-based algorithm to assist nephropathy diagnosis. This paper provides an
alternative solution for integrating multi-stained visual cues for glomerulus
classification. Our approach, named generator-to-classifier (G2C), is a
two-stage framework. Given an input image from a specified stain, several
generators are first applied to estimate its appearances in other staining
methods, and a classifier follows to combine visual cues from different stains
for prediction (whether it is pathological, or which type of pathology it has).
We optimize these two stages in a joint manner. To provide a reasonable
initialization, we pre-train the generators in an unlabeled reference set under
an unpaired image-to-image translation task, and then fine-tune them together
with the classifier. We conduct experiments on a glomerulus type classification
dataset collected by ourselves (there are no publicly available datasets for
this purpose). Although joint optimization slightly harms the authenticity of
the generated patches, it boosts classification performance, suggesting more
effective visual cues are extracted in an automatic way. We also transfer our
model to a public dataset for breast cancer classification, and outperform the
state-of-the-arts significantly.Comment: Accepted by AAAI 201
Breast Cancer Immunohistochemical Image Generation: a Benchmark Dataset and Challenge Review
For invasive breast cancer, immunohistochemical (IHC) techniques are often
used to detect the expression level of human epidermal growth factor receptor-2
(HER2) in breast tissue to formulate a precise treatment plan. From the
perspective of saving manpower, material and time costs, directly generating
IHC-stained images from hematoxylin and eosin (H&E) stained images is a
valuable research direction. Therefore, we held the breast cancer
immunohistochemical image generation challenge, aiming to explore novel ideas
of deep learning technology in pathological image generation and promote
research in this field. The challenge provided registered H&E and IHC-stained
image pairs, and participants were required to use these images to train a
model that can directly generate IHC-stained images from corresponding
H&E-stained images. We selected and reviewed the five highest-ranking methods
based on their PSNR and SSIM metrics, while also providing overviews of the
corresponding pipelines and implementations. In this paper, we further analyze
the current limitations in the field of breast cancer immunohistochemical image
generation and forecast the future development of this field. We hope that the
released dataset and the challenge will inspire more scholars to jointly study
higher-quality IHC-stained image generation.Comment: 13 pages, 11 figures, 2table
Improved Breast Cancer Diagnosis through Transfer Learning on Hematoxylin and Eosin Stained Histology Images
Breast cancer is one of the leading causes of death for women worldwide.
Early screening is essential for early identification, but the chance of
survival declines as the cancer progresses into advanced stages. For this
study, the most recent BRACS dataset of histological (H\&E) stained images was
used to classify breast cancer tumours, which contains both the whole-slide
images (WSI) and region-of-interest (ROI) images, however, for our study we
have considered ROI images. We have experimented using different pre-trained
deep learning models, such as Xception, EfficientNet, ResNet50, and
InceptionResNet, pre-trained on the ImageNet weights. We pre-processed the
BRACS ROI along with image augmentation, upsampling, and dataset split
strategies. For the default dataset split, the best results were obtained by
ResNet50 achieving 66\% f1-score. For the custom dataset split, the best
results were obtained by performing upsampling and image augmentation which
results in 96.2\% f1-score. Our second approach also reduced the number of
false positive and false negative classifications to less than 3\% for each
class. We believe that our study significantly impacts the early diagnosis and
identification of breast cancer tumors and their subtypes, especially atypical
and malignant tumors, thus improving patient outcomes and reducing patient
mortality rates. Overall, this study has primarily focused on identifying seven
(7) breast cancer tumor subtypes, and we believe that the experimental models
can be fine-tuned further to generalize over previous breast cancer histology
datasets as well.Comment: 11 pages, 4 figure
Deep Learning in Breast Cancer Imaging: A Decade of Progress and Future Directions
Breast cancer has reached the highest incidence rate worldwide among all
malignancies since 2020. Breast imaging plays a significant role in early
diagnosis and intervention to improve the outcome of breast cancer patients. In
the past decade, deep learning has shown remarkable progress in breast cancer
imaging analysis, holding great promise in interpreting the rich information
and complex context of breast imaging modalities. Considering the rapid
improvement in the deep learning technology and the increasing severity of
breast cancer, it is critical to summarize past progress and identify future
challenges to be addressed. In this paper, we provide an extensive survey of
deep learning-based breast cancer imaging research, covering studies on
mammogram, ultrasound, magnetic resonance imaging, and digital pathology images
over the past decade. The major deep learning methods, publicly available
datasets, and applications on imaging-based screening, diagnosis, treatment
response prediction, and prognosis are described in detail. Drawn from the
findings of this survey, we present a comprehensive discussion of the
challenges and potential avenues for future research in deep learning-based
breast cancer imaging.Comment: Survey, 41 page
Representation learning for breast cancer lesion detection
Breast Cancer (BC) is the second type of cancer with a higher incidence in women, it is responsible for the death of hundreds of thousands of women every year. However, when detected in the early stages of the disease, treatment methods have proven to be very effective in increasing life expectancy and, in many cases, patients fully recover. Several medical image modalities, such as MG – Mammography (X-Rays), US - Ultrasound, CT - Computer Tomography, MRI - Magnetic Resonance Imaging, and Tomosynthesis have been explored to support radiologists/physicians in clinical decision-making work- flows for the detection and diagnosis of BC. MG is the imaging modality more used at the worldwide level, however, recent research results have demonstrated that breast MRI is more sensitive than mam- mography to find pathological lesions, and it is not limited/affected by breast density issues. Therefore, it is currently a trend to introduce MRI-based breast assessment into clinical workflows (screening and diagnosis), but when compared to MG the workload of radiologists/physicians increases, MRI assess- ment is a more time-consuming task, and its effectiveness is affected not only by the variety of morpho- logical characteristics of each specific tumor phenotype and its origin but also by human fatigue. Com- puter-Aided Detection (CADe) methods have been widely explored primarily in mammography screen- ing tasks, but it remains an unsolved problem in breast MRI settings.
This work aims to explore and validate BC detection models using Machine (Deep) Learning algorithms. As the main contribution, we have developed and validated an innovative method that improves the “breast MRI preprocessing phase” to select the patient’s image slices and bounding boxes representing pathological lesions. With this, it is possible to build a more robust training dataset to feed the deep learning models, reducing the computation time and the dimension of the dataset, and more importantly, to identify with high accuracy the specific regions (bounding boxes) for each of the patient images, in which a possible pathological lesion (tumor) has been identified. In experimental settings using a fully annotated (released for public domain) dataset comprising a total of 922 MRI-based BC patient cases, we have achieved, as the most accurate trained model, an accuracy rate of 97.83%, and subsequently, applying a ten-fold cross-validation method, a mean accuracy on the trained models of 94.46% and an associated standard deviation of 2.43%.O cancro da mama (CdM) é o segundo tipo de cancro com maior incidência nas mulheres. É respon- sável pela morte de centenas de milhares de mulheres todos os anos. Contudo, quando detetado nas fases iniciais da doença, os métodos de tratamento provaram ser muito eficazes aumentando a espe- rança de vida e, em muitos casos, os pacientes recuperam totalmente. Têm sido exploradas várias modalidades de imagem médica, tais como MG - Mamografia (Raios-X), US - Ultra-som, CT - Tomo- grafia Computadorizada, MRI - Ressonância Magnética e Tomossíntese, para apoiar radiologistas nos fluxos de trabalho clínico para a deteção e diagnóstico do CdM. A MG é a modalidade de imagem mais utilizada a nível mundial, contudo, resultados de pesquisas recentes demonstraram que o MRI é mais sensível do que a mamografia para encontrar lesões patológicas e, também, não é limitada ou afetada por questões de densidade mamária. Consequentemente, atualmente é uma tendência introduzir a avaliação mamográfica baseada em MRI nos fluxos de trabalho clínico - rastreio e diagnóstico -, mas quando comparada com a MG, a carga de trabalho dos radiologistas aumenta. A avaliação por MRI é uma tarefa mais demorada, e a sua eficácia é afetada não só pela variedade de características morfo- lógicas e origem de cada fenótipo tumoral específico, mas, também pela fadiga humana. Os métodos de deteção assistida por computador (CADe) têm sido amplamente explorados principalmente em ta- refas de rastreio mamográfico, mas continua a ser um problema por resolver em ambientes de resso- nância magnética mamária.
Este trabalho visa explorar e validar modelos de deteção de CdM usando algoritmos de Machine (Deep) Learning. Como contributo principal, desenvolvemos e validámos um método inovador que me- lhora a "fase de pré-processamento das imagens de ressonância magnética mamária" para selecionar as fatias de imagem do paciente e as respetivas caixas de contorno que representam as lesões pato- lógicas. Com isto, é possível construir um conjunto de dados de treino mais robusto para alimentar os modelos de deep learning, reduzir o tempo de computação, reduzir a dimensão do conjunto de dados e, mais importante, para identificar com alta precisão as regiões específicas para cada uma das ima- gens do paciente nas quais foi identificada uma possível lesão patológica (tumor). Os resultados expe- rimentais, num conjunto de imagens de ressonância magnética de domínio público totalmente anotado com 922 casos de doentes com CdM, mostram no melhor modelo uma taxa de exatidão de 97.83%. Foi aplicado um método de validação cruzada de 10 folds do qual resultou uma exatidão média de 94,46% com um desvio padrão de 2,43% nos modelos treinados
Magnification-independent Histopathological Image Classification with Similarity-based Multi-scale Embeddings
The classification of histopathological images is of great value in both
cancer diagnosis and pathological studies. However, multiple reasons, such as
variations caused by magnification factors and class imbalance, make it a
challenging task where conventional methods that learn from image-label
datasets perform unsatisfactorily in many cases. We observe that tumours of the
same class often share common morphological patterns. To exploit this fact, we
propose an approach that learns similarity-based multi-scale embeddings (SMSE)
for magnification-independent histopathological image classification. In
particular, a pair loss and a triplet loss are leveraged to learn
similarity-based embeddings from image pairs or image triplets. The learned
embeddings provide accurate measurements of similarities between images, which
are regarded as a more effective form of representation for histopathological
morphology than normal image features. Furthermore, in order to ensure the
generated models are magnification-independent, images acquired at different
magnification factors are simultaneously fed to networks during training for
learning multi-scale embeddings. In addition to the SMSE, to eliminate the
impact of class imbalance, instead of using the hard sample mining strategy
that intuitively discards some easy samples, we introduce a new reinforced
focal loss to simultaneously punish hard misclassified samples while
suppressing easy well-classified samples. Experimental results show that the
SMSE improves the performance for histopathological image classification tasks
for both breast and liver cancers by a large margin compared to previous
methods. In particular, the SMSE achieves the best performance on the BreakHis
benchmark with an improvement ranging from 5% to 18% compared to previous
methods using traditional features
Advancing early leukemia diagnostics: a comprehensive study incorporating image processing and transfer learning.
Disease recognition has been revolutionized by autonomous systems in the rapidly developing field of medical technology. A crucial aspect of diagnosis involves the visual assessment and enumeration of white blood cells in microscopic peripheral blood smears. This practice yields invaluable insights into a patient's health, enabling the identification of conditions of blood malignancies such as leukemia. Early identification of leukemia subtypes is paramount for tailoring appropriate therapeutic interventions and enhancing patient survival rates. However, traditional diagnostic techniques, which depend on visual assessment, are arbitrary, laborious, and prone to errors. The advent of ML technologies offers a promising avenue for more accurate and efficient leukemia classification. In this study, we introduced a novel approach to leukemia classification by integrating advanced image processing, diverse dataset utilization, and sophisticated feature extraction techniques, coupled with the development of TL models. Focused on improving accuracy of previous studies, our approach utilized Kaggle datasets for binary and multiclass classifications. Extensive image processing involved a novel LoGMH method, complemented by diverse augmentation techniques. Feature extraction employed DCNN, with subsequent utilization of extracted features to train various ML and TL models. Rigorous evaluation using traditional metrics revealed Inception-ResNet's superior performance, surpassing other models with F1 scores of 96.07% and 95.89% for binary and multiclass classification, respectively. Our results notably surpass previous research, particularly in cases involving a higher number of classes. These findings promise to influence clinical decision support systems, guide future research, and potentially revolutionize cancer diagnostics beyond leukemia, impacting broader medical imaging and oncology domains
Efficient breast cancer classification network with dual squeeze and excitation in histopathological images.
Medical image analysis methods for mammograms, ultrasound, and magnetic resonance imaging (MRI) cannot provide the underline features on the cellular level to understand the cancer microenvironment which makes them unsuitable for breast cancer subtype classification study. In this paper, we propose a convolutional neural network (CNN)-based breast cancer classification method for hematoxylin and eosin (H&E) whole slide images (WSIs). The proposed method incorporates fused mobile inverted bottleneck convolutions (FMB-Conv) and mobile inverted bottleneck convolutions (MBConv) with a dual squeeze and excitation (DSE) network to accurately classify breast cancer tissue into binary (benign and malignant) and eight subtypes using histopathology images. For that, a pre-trained EfficientNetV2 network is used as a backbone with a modified DSE block that combines the spatial and channel-wise squeeze and excitation layers to highlight important low-level and high-level abstract features. Our method outperformed ResNet101, InceptionResNetV2, and EfficientNetV2 networks on the publicly available BreakHis dataset for the binary and multi-class breast cancer classification in terms of precision, recall, and F1-score on multiple magnification levels
- …