313 research outputs found
Weakly Supervised Deep Learning for Thoracic Disease Classification and Localization on Chest X-rays
Chest X-rays is one of the most commonly available and affordable
radiological examinations in clinical practice. While detecting thoracic
diseases on chest X-rays is still a challenging task for machine intelligence,
due to 1) the highly varied appearance of lesion areas on X-rays from patients
of different thoracic disease and 2) the shortage of accurate pixel-level
annotations by radiologists for model training. Existing machine learning
methods are unable to deal with the challenge that thoracic diseases usually
happen in localized disease-specific areas. In this article, we propose a
weakly supervised deep learning framework equipped with squeeze-and-excitation
blocks, multi-map transfer, and max-min pooling for classifying thoracic
diseases as well as localizing suspicious lesion regions. The comprehensive
experiments and discussions are performed on the ChestX-ray14 dataset. Both
numerical and visual results have demonstrated the effectiveness of the
proposed model and its better performance against the state-of-the-art
pipelines.Comment: 10 pages. Accepted by the ACM BCB 201
Automated Detection and Forecasting of COVID-19 using Deep Learning Techniques: A Review
Coronavirus, or COVID-19, is a hazardous disease that has endangered the
health of many people around the world by directly affecting the lungs.
COVID-19 is a medium-sized, coated virus with a single-stranded RNA. This virus
has one of the largest RNA genomes and is approximately 120 nm. The X-Ray and
computed tomography (CT) imaging modalities are widely used to obtain a fast
and accurate medical diagnosis. Identifying COVID-19 from these medical images
is extremely challenging as it is time-consuming, demanding, and prone to human
errors. Hence, artificial intelligence (AI) methodologies can be used to obtain
consistent high performance. Among the AI methodologies, deep learning (DL)
networks have gained much popularity compared to traditional machine learning
(ML) methods. Unlike ML techniques, all stages of feature extraction, feature
selection, and classification are accomplished automatically in DL models. In
this paper, a complete survey of studies on the application of DL techniques
for COVID-19 diagnostic and automated segmentation of lungs is discussed,
concentrating on works that used X-Ray and CT images. Additionally, a review of
papers on the forecasting of coronavirus prevalence in different parts of the
world with DL techniques is presented. Lastly, the challenges faced in the
automated detection of COVID-19 using DL techniques and directions for future
research are discussed
Is attention all you need in medical image analysis? A review
Medical imaging is a key component in clinical diagnosis, treatment planning
and clinical trial design, accounting for almost 90% of all healthcare data.
CNNs achieved performance gains in medical image analysis (MIA) over the last
years. CNNs can efficiently model local pixel interactions and be trained on
small-scale MI data. The main disadvantage of typical CNN models is that they
ignore global pixel relationships within images, which limits their
generalisation ability to understand out-of-distribution data with different
'global' information. The recent progress of Artificial Intelligence gave rise
to Transformers, which can learn global relationships from data. However, full
Transformer models need to be trained on large-scale data and involve
tremendous computational complexity. Attention and Transformer compartments
(Transf/Attention) which can well maintain properties for modelling global
relationships, have been proposed as lighter alternatives of full Transformers.
Recently, there is an increasing trend to co-pollinate complementary
local-global properties from CNN and Transf/Attention architectures, which led
to a new era of hybrid models. The past years have witnessed substantial growth
in hybrid CNN-Transf/Attention models across diverse MIA problems. In this
systematic review, we survey existing hybrid CNN-Transf/Attention models,
review and unravel key architectural designs, analyse breakthroughs, and
evaluate current and future opportunities as well as challenges. We also
introduced a comprehensive analysis framework on generalisation opportunities
of scientific and clinical impact, based on which new data-driven domain
generalisation and adaptation methods can be stimulated
HydraViT: Adaptive Multi-Branch Transformer for Multi-Label Disease Classification from Chest X-ray Images
Chest X-ray is an essential diagnostic tool in the identification of chest
diseases given its high sensitivity to pathological abnormalities in the lungs.
However, image-driven diagnosis is still challenging due to heterogeneity in
size and location of pathology, as well as visual similarities and
co-occurrence of separate pathology. Since disease-related regions often occupy
a relatively small portion of diagnostic images, classification models based on
traditional convolutional neural networks (CNNs) are adversely affected given
their locality bias. While CNNs were previously augmented with attention maps
or spatial masks to guide focus on potentially critical regions, learning
localization guidance under heterogeneity in the spatial distribution of
pathology is challenging. To improve multi-label classification performance,
here we propose a novel method, HydraViT, that synergistically combines a
transformer backbone with a multi-branch output module with learned weighting.
The transformer backbone enhances sensitivity to long-range context in X-ray
images, while using the self-attention mechanism to adaptively focus on
task-critical regions. The multi-branch output module dedicates an independent
branch to each disease label to attain robust learning across separate disease
classes, along with an aggregated branch across labels to maintain sensitivity
to co-occurrence relationships among pathology. Experiments demonstrate that,
on average, HydraViT outperforms competing attention-guided methods by 1.2%,
region-guided methods by 1.4%, and semantic-guided methods by 1.0% in
multi-label classification performance
- …