28 research outputs found
VORTEX: Physics-Driven Data Augmentations Using Consistency Training for Robust Accelerated MRI Reconstruction
Deep neural networks have enabled improved image quality and fast inference
times for various inverse problems, including accelerated magnetic resonance
imaging (MRI) reconstruction. However, such models require a large number of
fully-sampled ground truth datasets, which are difficult to curate, and are
sensitive to distribution drifts. In this work, we propose applying
physics-driven data augmentations for consistency training that leverage our
domain knowledge of the forward MRI data acquisition process and MRI physics to
achieve improved label efficiency and robustness to clinically-relevant
distribution drifts. Our approach, termed VORTEX, (1) demonstrates strong
improvements over supervised baselines with and without data augmentation in
robustness to signal-to-noise ratio change and motion corruption in
data-limited regimes; (2) considerably outperforms state-of-the-art purely
image-based data augmentation techniques and self-supervised reconstruction
methods on both in-distribution and out-of-distribution data; and (3) enables
composing heterogeneous image-based and physics-driven data augmentations. Our
code is available at https://github.com/ad12/meddlr.Comment: Accepted to MIDL 202
The International Workshop on Osteoarthritis Imaging Knee MRI Segmentation Challenge: A Multi-Institute Evaluation and Analysis Framework on a Standardized Dataset
Purpose: To organize a knee MRI segmentation challenge for characterizing the
semantic and clinical efficacy of automatic segmentation methods relevant for
monitoring osteoarthritis progression.
Methods: A dataset partition consisting of 3D knee MRI from 88 subjects at
two timepoints with ground-truth articular (femoral, tibial, patellar)
cartilage and meniscus segmentations was standardized. Challenge submissions
and a majority-vote ensemble were evaluated using Dice score, average symmetric
surface distance, volumetric overlap error, and coefficient of variation on a
hold-out test set. Similarities in network segmentations were evaluated using
pairwise Dice correlations. Articular cartilage thickness was computed per-scan
and longitudinally. Correlation between thickness error and segmentation
metrics was measured using Pearson's coefficient. Two empirical upper bounds
for ensemble performance were computed using combinations of model outputs that
consolidated true positives and true negatives.
Results: Six teams (T1-T6) submitted entries for the challenge. No
significant differences were observed across all segmentation metrics for all
tissues (p=1.0) among the four top-performing networks (T2, T3, T4, T6). Dice
correlations between network pairs were high (>0.85). Per-scan thickness errors
were negligible among T1-T4 (p=0.99) and longitudinal changes showed minimal
bias (<0.03mm). Low correlations (<0.41) were observed between segmentation
metrics and thickness error. The majority-vote ensemble was comparable to top
performing networks (p=1.0). Empirical upper bound performances were similar
for both combinations (p=1.0).
Conclusion: Diverse networks learned to segment the knee similarly where high
segmentation accuracy did not correlate to cartilage thickness accuracy. Voting
ensembles did not outperform individual networks but may help regularize
individual models.Comment: Submitted to Radiology: Artificial Intelligence; Fixed typo
Noise2Recon: Enabling Joint MRI Reconstruction and Denoising with Semi-Supervised and Self-Supervised Learning
Deep learning (DL) has shown promise for faster, high quality accelerated MRI
reconstruction. However, supervised DL methods depend on extensive amounts of
fully-sampled (labeled) data and are sensitive to out-of-distribution (OOD)
shifts, particularly low signal-to-noise ratio (SNR) acquisitions. To alleviate
this challenge, we propose Noise2Recon, a model-agnostic, consistency training
method for joint MRI reconstruction and denoising that can use both
fully-sampled (labeled) and undersampled (unlabeled) scans in semi-supervised
and self-supervised settings. With limited or no labeled training data,
Noise2Recon outperforms compressed sensing and deep learning baselines,
including supervised networks, augmentation-based training, fine-tuned
denoisers, and self-supervised methods, and matches performance of supervised
models, which were trained with 14x more fully-sampled scans. Noise2Recon also
outperforms all baselines, including state-of-the-art fine-tuning and
augmentation techniques, among low-SNR scans and when generalizing to other OOD
factors, such as changes in acceleration factors and different datasets.
Augmentation extent and loss weighting hyperparameters had negligible impact on
Noise2Recon compared to supervised methods, which may indicate increased
training stability. Our code is available at https://github.com/ad12/meddlr
RadAdapt: Radiology Report Summarization via Lightweight Domain Adaptation of Large Language Models
We systematically investigate lightweight strategies to adapt large language
models (LLMs) for the task of radiology report summarization (RRS).
Specifically, we focus on domain adaptation via pretraining (on natural
language, biomedical text, and clinical text) and via prompting (zero-shot,
in-context learning) or parameter-efficient fine-tuning (prefix tuning, LoRA).
Our results on the MIMIC-III dataset consistently demonstrate best performance
by maximally adapting to the task via pretraining on clinical text and
parameter-efficient fine-tuning on RRS examples. Importantly, this method
fine-tunes a mere 0.32% of parameters throughout the model, in contrast to
end-to-end fine-tuning (100% of parameters). Additionally, we study the effect
of in-context examples and out-of-distribution (OOD) training before concluding
with a radiologist reader study and qualitative analysis. Our findings
highlight the importance of domain adaptation in RRS and provide valuable
insights toward developing effective natural language processing solutions for
clinical tasks.Comment: 12 pages, 9 figure
Clinical Text Summarization: Adapting Large Language Models Can Outperform Human Experts
Sifting through vast textual data and summarizing key information imposes a
substantial burden on how clinicians allocate their time. Although large
language models (LLMs) have shown immense promise in natural language
processing (NLP) tasks, their efficacy across diverse clinical summarization
tasks has not yet been rigorously examined. In this work, we employ domain
adaptation methods on eight LLMs, spanning six datasets and four distinct
summarization tasks: radiology reports, patient questions, progress notes, and
doctor-patient dialogue. Our thorough quantitative assessment reveals
trade-offs between models and adaptation methods in addition to instances where
recent advances in LLMs may not lead to improved results. Further, in a
clinical reader study with six physicians, we depict that summaries from the
best adapted LLM are preferable to human summaries in terms of completeness and
correctness. Our ensuing qualitative analysis delineates mutual challenges
faced by both LLMs and human experts. Lastly, we correlate traditional
quantitative NLP metrics with reader study scores to enhance our understanding
of how these metrics align with physician preferences. Our research marks the
first evidence of LLMs outperforming human experts in clinical text
summarization across multiple tasks. This implies that integrating LLMs into
clinical workflows could alleviate documentation burden, empowering clinicians
to focus more on personalized patient care and other irreplaceable human
aspects of medicine.Comment: 23 pages, 22 figure
Self-supervised learning for medical image classification: a systematic review and implementation guidelines
Abstract Advancements in deep learning and computer vision provide promising solutions for medical image analysis, potentially improving healthcare and patient outcomes. However, the prevailing paradigm of training deep learning models requires large quantities of labeled training data, which is both time-consuming and cost-prohibitive to curate for medical images. Self-supervised learning has the potential to make significant contributions to the development of robust medical imaging models through its ability to learn useful insights from copious medical datasets without labels. In this review, we provide consistent descriptions of different self-supervised learning strategies and compose a systematic review of papers published between 2012 and 2022 on PubMed, Scopus, and ArXiv that applied self-supervised learning to medical imaging classification. We screened a total of 412 relevant studies and included 79 papers for data extraction and analysis. With this comprehensive effort, we synthesize the collective knowledge of prior work and provide implementation guidelines for future researchers interested in applying self-supervised learning to their development of medical imaging classification models