125,295 research outputs found
Deep Semantic Segmentation of Natural and Medical Images: A Review
The semantic image segmentation task consists of classifying each pixel of an
image into an instance, where each instance corresponds to a class. This task
is a part of the concept of scene understanding or better explaining the global
context of an image. In the medical image analysis domain, image segmentation
can be used for image-guided interventions, radiotherapy, or improved
radiological diagnostics. In this review, we categorize the leading deep
learning-based medical and non-medical image segmentation solutions into six
main groups of deep architectural, data synthesis-based, loss function-based,
sequenced models, weakly supervised, and multi-task methods and provide a
comprehensive review of the contributions in each of these groups. Further, for
each group, we analyze each variant of these groups and discuss the limitations
of the current approaches and present potential future research directions for
semantic image segmentation.Comment: 45 pages, 16 figures. Accepted for publication in Springer Artificial
Intelligence Revie
Deployment of Artificial Intelligence in Real-World Practice: Opportunity and Challenge.
Artificial intelligence has rapidly evolved from the experimental phase to the implementation phase in many image-driven clinical disciplines, including ophthalmology. A combination of the increasing availability of large datasets and computing power with revolutionary progress in deep learning has created unprecedented opportunities for major breakthrough improvements in the performance and accuracy of automated diagnoses that primarily focus on image recognition and feature detection. Such an automated disease classification would significantly improve the accessibility, efficiency, and cost-effectiveness of eye care systems where it is less dependent on human input, potentially enabling diagnosis to be cheaper, quicker, and more consistent. Although this technology will have a profound impact on clinical flow and practice patterns sooner or later, translating such a technology into clinical practice is challenging and requires similar levels of accountability and effectiveness as any new medication or medical device due to the potential problems of bias, and ethical, medical, and legal issues that might arise. The objective of this review is to summarize the opportunities and challenges of this transition and to facilitate the integration of artificial intelligence (AI) into routine clinical practice based on our best understanding and experience in this area
Robotic Ultrasound Imaging: State-of-the-Art and Future Perspectives
Ultrasound (US) is one of the most widely used modalities for clinical
intervention and diagnosis due to the merits of providing non-invasive,
radiation-free, and real-time images. However, free-hand US examinations are
highly operator-dependent. Robotic US System (RUSS) aims at overcoming this
shortcoming by offering reproducibility, while also aiming at improving
dexterity, and intelligent anatomy and disease-aware imaging. In addition to
enhancing diagnostic outcomes, RUSS also holds the potential to provide medical
interventions for populations suffering from the shortage of experienced
sonographers. In this paper, we categorize RUSS as teleoperated or autonomous.
Regarding teleoperated RUSS, we summarize their technical developments, and
clinical evaluations, respectively. This survey then focuses on the review of
recent work on autonomous robotic US imaging. We demonstrate that machine
learning and artificial intelligence present the key techniques, which enable
intelligent patient and process-specific, motion and deformation-aware robotic
image acquisition. We also show that the research on artificial intelligence
for autonomous RUSS has directed the research community toward understanding
and modeling expert sonographers' semantic reasoning and action. Here, we call
this process, the recovery of the "language of sonography". This side result of
research on autonomous robotic US acquisitions could be considered as valuable
and essential as the progress made in the robotic US examination itself. This
article will provide both engineers and clinicians with a comprehensive
understanding of RUSS by surveying underlying techniques.Comment: Accepted by Medical Image Analysi
Utilizing Segment Anything Model For Assessing Localization of GRAD-CAM in Medical Imaging
The introduction of saliency map algorithms as an approach for assessing the
interoperability of images has allowed for a deeper understanding of current
black-box models with Artificial Intelligence. Their rise in popularity has led
to these algorithms being applied in multiple fields, including medical
imaging. With a classification task as important as those in the medical
domain, a need for rigorous testing of their capabilities arises. Current works
examine capabilities through assessing the localization of saliency maps upon
medical abnormalities within an image, through comparisons with human
annotations. We propose utilizing Segment Anything Model (SAM) to both further
the accuracy of such existing metrics, while also generalizing beyond the need
for human annotations. Our results show both high degrees of similarity to
existing metrics while also highlighting the capabilities of this methodology
to beyond human-annotation. Furthermore, we explore the applications (and
challenges) of SAM within the medical domain, including image pre-processing
before segmenting, natural language proposals to SAM in the form of CLIP-SAM,
and SAM accuracy across multiple medical imaging datasets.Comment: 11 pages, 14 figures, 1 tabl
Enhancing Medical Image Segmentation: Optimizing Cross-Entropy Weights and Post-Processing with Autoencoders
The task of medical image segmentation presents unique challenges,
necessitating both localized and holistic semantic understanding to accurately
delineate areas of interest, such as critical tissues or aberrant features.
This complexity is heightened in medical image segmentation due to the high
degree of inter-class similarities, intra-class variations, and possible image
obfuscation. The segmentation task further diversifies when considering the
study of histopathology slides for autoimmune diseases like dermatomyositis.
The analysis of cell inflammation and interaction in these cases has been less
studied due to constraints in data acquisition pipelines. Despite the
progressive strides in medical science, we lack a comprehensive collection of
autoimmune diseases. As autoimmune diseases globally escalate in prevalence and
exhibit associations with COVID-19, their study becomes increasingly essential.
While there is existing research that integrates artificial intelligence in the
analysis of various autoimmune diseases, the exploration of dermatomyositis
remains relatively underrepresented. In this paper, we present a deep-learning
approach tailored for Medical image segmentation. Our proposed method
outperforms the current state-of-the-art techniques by an average of 12.26% for
U-Net and 12.04% for U-Net++ across the ResNet family of encoders on the
dermatomyositis dataset. Furthermore, we probe the importance of optimizing
loss function weights and benchmark our methodology on three challenging
medical image segmentation tasksComment: Accepted at ICCV CVAMD 202
Evaluating Explainable AI on a Multi-Modal Medical Imaging Task: Can Existing Algorithms Fulfill Clinical Requirements?
Being able to explain the prediction to clinical end-users is a necessity to
leverage the power of artificial intelligence (AI) models for clinical decision
support. For medical images, a feature attribution map, or heatmap, is the most
common form of explanation that highlights important features for AI models'
prediction. However, it is unknown how well heatmaps perform on explaining
decisions on multi-modal medical images, where each image modality or channel
visualizes distinct clinical information of the same underlying biomedical
phenomenon. Understanding such modality-dependent features is essential for
clinical users' interpretation of AI decisions. To tackle this clinically
important but technically ignored problem, we propose the modality-specific
feature importance (MSFI) metric. It encodes clinical image and explanation
interpretation patterns of modality prioritization and modality-specific
feature localization. We conduct a clinical requirement-grounded, systematic
evaluation using computational methods and a clinician user study. Results show
that the examined 16 heatmap algorithms failed to fulfill clinical requirements
to correctly indicate AI model decision process or decision quality. The
evaluation and MSFI metric can guide the design and selection of XAI algorithms
to meet clinical requirements on multi-modal explanation.Comment: AAAI 202
- …