28 research outputs found
Few-shot medical image classification with simple shape and texture text descriptors using vision-language models
In this work, we investigate the usefulness of vision-language models (VLMs)
and large language models for binary few-shot classification of medical images.
We utilize the GPT-4 model to generate text descriptors that encapsulate the
shape and texture characteristics of objects in medical images. Subsequently,
these GPT-4 generated descriptors, alongside VLMs pre-trained on natural
images, are employed to classify chest X-rays and breast ultrasound images. Our
results indicate that few-shot classification of medical images using VLMs and
GPT-4 generated descriptors is a viable approach. However, accurate
classification requires to exclude certain descriptors from the calculations of
the classification scores. Moreover, we assess the ability of VLMs to evaluate
shape features in breast mass ultrasound images. We further investigate the
degree of variability among the sets of text descriptors produced by GPT-4. Our
work provides several important insights about the application of VLMs for
medical image analysis.Comment: 13 pages, 5 figure
Implicit neural representations for joint decomposition and registration of gene expression images in the marmoset brain
We propose a novel image registration method based on implicit neural
representations that addresses the challenging problem of registering a pair of
brain images with similar anatomical structures, but where one image contains
additional features or artifacts that are not present in the other image. To
demonstrate its effectiveness, we use 2D microscopy
hybridization gene expression images of the marmoset brain. Accurately
quantifying gene expression requires image registration to a brain template,
which is difficult due to the diversity of patterns causing variations in
visible anatomical brain structures. Our approach uses implicit networks in
combination with an image exclusion loss to jointly perform the registration
and decompose the image into a support and residual image. The support image
aligns well with the template, while the residual image captures individual
image characteristics that diverge from the template. In experiments, our
method provided excellent results and outperformed other registration
techniques.Comment: 11 page
Quantitative Ultrasound and B-mode Image Texture Features Correlate with Collagen and Myelin Content in Human Ulnar Nerve Fascicles
We investigate the usefulness of quantitative ultrasound (QUS) and B-mode
texture features for characterization of ulnar nerve fascicles. Ultrasound data
were acquired from cadaveric specimens using a nominal 30 MHz probe. Next, the
nerves were extracted to prepare histology sections. 85 fascicles were matched
between the B-mode images and the histology sections. For each fascicle image,
we selected an intra-fascicular region of interest. We used histology sections
to determine features related to the concentration of collagen and myelin, and
ultrasound data to calculate backscatter coefficient (-24.89 dB 8.31),
attenuation coefficient (0.92 db/cm-MHz 0.04), Nakagami parameter (1.01
0.18) and entropy (6.92 0.83), as well as B-mode texture features
obtained via the gray level co-occurrence matrix algorithm. Significant
Spearman's rank correlations between the combined collagen and myelin
concentrations were obtained for the backscatter coefficient (R=-0.68), entropy
(R=-0.51), and for several texture features. Our study demonstrates that QUS
may potentially provide information on structural components of nerve
fascicles
PatchMorph: A Stochastic Deep Learning Approach for Unsupervised 3D Brain Image Registration with Small Patches
We introduce "PatchMorph," an new stochastic deep learning algorithm tailored
for unsupervised 3D brain image registration. Unlike other methods, our method
uses compact patches of a constant small size to derive solutions that can
combine global transformations with local deformations. This approach minimizes
the memory footprint of the GPU during training, but also enables us to operate
on numerous amounts of randomly overlapping small patches during inference to
mitigate image and patch boundary problems. PatchMorph adeptly handles world
coordinate transformations between two input images, accommodating variances in
attributes such as spacing, array sizes, and orientations. The spatial
resolution of patches transitions from coarse to fine, addressing both global
and local attributes essential for aligning the images. Each patch offers a
unique perspective, together converging towards a comprehensive solution.
Experiments on human T1 MRI brain images and marmoset brain images from serial
2-photon tomography affirm PatchMorph's superior performance.Comment: This work has been submitted to the IEEE for possible publication.
Copyright may be transferred without notice, after which this version may no
longer be accessibl
BUS-Set:A benchmark for quantitative evaluation of breast ultrasound segmentation networks with public datasets
Purpose: BUS-Set is a reproducible benchmark for breast ultrasound (BUS) lesion segmentation, comprising of publicly available images with the aim of improving future comparisons between machine learning models within the field of BUS. Method: Four publicly available datasets were compiled creating an overall set of 1154 BUS images, from five different scanner types. Full dataset details have been provided, which include clinical labels and detailed annotations. Furthermore, nine state-of-the-art deep learning architectures were selected to form the initial benchmark segmentation result, tested using five-fold cross-validation and MANOVA/ANOVA with Tukey statistical significance test with a threshold of 0.01. Additional evaluation of these architectures was conducted, exploring possible training bias, and lesion size and type effects. Results: Of the nine state-of-the-art benchmarked architectures, Mask R-CNN obtained the highest overall results, with the following mean metric scores: Dice score of 0.851, intersection over union of 0.786 and pixel accuracy of 0.975. MANOVA/ANOVA and Tukey test results showed Mask R-CNN to be statistically significant better compared to all other benchmarked models with a p-value >0.01. Moreover, Mask R-CNN achieved the highest mean Dice score of 0.839 on an additional 16 image dataset, that contained multiple lesions per image. Further analysis on regions of interest was conducted, assessing Hamming distance, depth-to-width ratio (DWR), circularity, and elongation, which showed that the Mask R-CNN's segmentations maintained the most morphological features with correlation coefficients of 0.888, 0.532, 0.876 for DWR, circularity, and elongation, respectively. Based on the correlation coefficients, statistical test indicated that Mask R-CNN was only significantly different to Sk-U-Net. Conclusions: BUS-Set is a fully reproducible benchmark for BUS lesion segmentation obtained through the use of public datasets and GitHub. Of the state-of-the-art convolution neural network (CNN)-based architectures, Mask R-CNN achieved the highest performance overall, further analysis indicated that a training bias may have occurred due to the lesion size variation in the dataset. All dataset and architecture details are available at GitHub: https://github.com/corcor27/BUS-Set, which allows for a fully reproducible benchmark
Recommended from our members
Knee menisci segmentation and relaxometry of 3D ultrashort echo time cones MR imaging using attention U-Net with transfer learning.
PurposeTo develop a deep learning-based method for knee menisci segmentation in 3D ultrashort echo time (UTE) cones MR imaging, and to automatically determine MR relaxation times, namely the T1, T1ρ , and T2∗ parameters, which can be used to assess knee osteoarthritis (OA).MethodsWhole knee joint imaging was performed using 3D UTE cones sequences to collect data from 61 human subjects. Regions of interest (ROIs) were outlined by 2 experienced radiologists based on subtracted T1ρ -weighted MR images. Transfer learning was applied to develop 2D attention U-Net convolutional neural networks for the menisci segmentation based on each radiologist's ROIs separately. Dice scores were calculated to assess segmentation performance. Next, the T1, T1ρ , T2∗ relaxations, and ROI areas were determined for the manual and automatic segmentations, then compared.ResultsThe models developed using ROIs provided by 2 radiologists achieved high Dice scores of 0.860 and 0.833, while the radiologists' manual segmentations achieved a Dice score of 0.820. Linear correlation coefficients for the T1, T1ρ , and T2∗ relaxations calculated using the automatic and manual segmentations ranged between 0.90 and 0.97, and there were no associated differences between the estimated average meniscal relaxation parameters. The deep learning models achieved segmentation performance equivalent to the inter-observer variability of 2 radiologists.ConclusionThe proposed deep learning-based approach can be used to efficiently generate automatic segmentations and determine meniscal relaxations times. The method has the potential to help radiologists with the assessment of meniscal diseases, such as OA