43 research outputs found

    Self-Supervised Learning for Spinal MRIs

    Full text link
    A significant proportion of patients scanned in a clinical setting have follow-up scans. We show in this work that such longitudinal scans alone can be used as a form of 'free' self-supervision for training a deep network. We demonstrate this self-supervised learning for the case of T2-weighted sagittal lumbar Magnetic Resonance Images (MRIs). A Siamese convolutional neural network (CNN) is trained using two losses: (i) a contrastive loss on whether the scan is of the same person (i.e. longitudinal) or not, together with (ii) a classification loss on predicting the level of vertebral bodies. The performance of this pre-trained network is then assessed on a grading classification task. We experiment on a dataset of 1016 subjects, 423 possessing follow-up scans, with the end goal of learning the disc degeneration radiological gradings attached to the intervertebral discs. We show that the performance of the pre-trained CNN on the supervised classification task is (i) superior to that of a network trained from scratch; and (ii) requires far fewer annotated training samples to reach an equivalent performance to that of the network trained from scratch.Comment: 3rd Workshop on Deep Learning in Medical Image Analysi

    Vision-Language Modelling For Radiological Imaging and Reports In The Low Data Regime

    Full text link
    This paper explores training medical vision-language models (VLMs) -- where the visual and language inputs are embedded into a common space -- with a particular focus on scenarios where training data is limited, as is often the case in clinical datasets. We explore several candidate methods to improve low-data performance, including: (i) adapting generic pre-trained models to novel image and text domains (i.e. medical imaging and reports) via unimodal self-supervision; (ii) using local (e.g. GLoRIA) & global (e.g. InfoNCE) contrastive loss functions as well as a combination of the two; (iii) extra supervision during VLM training, via: (a) image- and text-only self-supervision, and (b) creating additional positive image-text pairs for training through augmentation and nearest-neighbour search. Using text-to-image retrieval as a benchmark, we evaluate the performance of these methods with variable sized training datasets of paired chest X-rays and radiological reports. Combined, they significantly improve retrieval compared to fine-tuning CLIP, roughly equivalent to training with the data. A similar pattern is found in the downstream task classification of CXR-related conditions with our method outperforming CLIP and also BioVIL, a strong CXR VLM benchmark, in the zero-shot and linear probing settings. We conclude with a set of recommendations for researchers aiming to train vision-language models on other medical imaging modalities when training data is scarce. To facilitate further research, we will make our code and models publicly available.Comment: Accepted to MIDL 202

    Identifying scoliosis in population-based cohorts:automation of a validated method based on total body dual energy X-ray absorptiometry scans

    Get PDF
    Scoliosis is a 3D-torsional rotation of the spine, but risk factors for initiation and progression are little understood. Research is hampered by lack of population-based research since radiographs cannot be performed on entire populations due to the relatively high levels of ionising radiation. Hence we have developed and validated a manual method for identifying scoliosis from total body dual energy X-ray absorptiometry (DXA) scans for research purposes. However, to allow full utilisation of population-based research cohorts, this needs to be automated. The purpose of this study was therefore to automate the identification of spinal curvature from total body DXA scans using machine learning techniques. To validate the automation, we assessed: (1) sensitivity, specificity and area under the receiver operator curve value (AUC) by comparison with 12,000 manually annotated images; (2) reliability by rerunning the automation on a subset of DXA scans repeated 2–6 weeks apart and calculating the kappa statistic; (3) validity by applying the automation to 5000 non-annotated images to assess associations with epidemiological variables. The final automated model had a sensitivity of 86.5%, specificity of 96.9% and an AUC of 0.80 (95%CI 0.74–0.87). There was almost perfect agreement of identification of those with scoliosis (kappa 0.90). Those with scoliosis identified by the automated model showed similar associations with gender, ethnicity, socioeconomic status, BMI and lean mass to previous literature. In conclusion, we have developed an accurate and valid automated method for identifying and quantifying spinal curvature from total body DXA scans

    Lung cancer prediction by Deep Learning to identify benign lung nodules

    Get PDF
    INTRODUCTION: Deep Learning has been proposed as promising tool to classify malignant nodules. Our aim was to retrospectively validate our Lung Cancer Prediction Convolutional Neural Network (LCP-CNN), which was trained on US screening data, on an independent dataset of indeterminate nodules in an European multicentre trial, to rule out benign nodules maintaining a high lung cancer sensitivity. METHODS: The LCP-CNN has been trained to generate a malignancy score for each nodule using CT data from the U.S. National Lung Screening Trial (NLST), and validated on CT scans containing 2106 nodules (205 lung cancers) detected in patients from from the Early Lung Cancer Diagnosis Using Artificial Intelligence and Big Data (LUCINDA) study, recruited from three tertiary referral centers in the UK, Germany and Netherlands. We pre-defined a benign nodule rule-out test, to identify benign nodules whilst maintaining a high sensitivity, by calculating thresholds on the malignancy score that achieve at least 99 % sensitivity on the NLST data. Overall performance per validation site was evaluated using Area-Under-the-ROC-Curve analysis (AUC). RESULTS: The overall AUC across the European centers was 94.5 % (95 %CI 92.6-96.1). With a high sensitivity of 99.0 %, malignancy could be ruled out in 22.1 % of the nodules, enabling 18.5 % of the patients to avoid follow-up scans. The two false-negative results both represented small typical carcinoids. CONCLUSION: The LCP-CNN, trained on participants with lung nodules from the US NLST dataset, showed excellent performance on identification of benign lung nodules in a multi-center external dataset, ruling out malignancy with high accuracy in about one fifth of the patients with 5-15 mm nodules

    Geometric and photometric invariant distinctive regions detection

    No full text
    In this paper, we present a number of enhancements to the Kadir/Brady salient region detector which result in a significant improvement in performance. The modifications we make include: stabilising the difference between consecutive scales when calculating the inter-scale saliency, a new sampling strategy using overlap of pixels, partial volume estimation and parzen windowing. Repeatability is used as the criterion for evaluating the performance of the algorithm. We observe the repeatability for distinctive regions selected from an image and from the same image after applying a particular transformation. The transformations we use include planar rotation, pixel translation, spatial scaling, and intensity shifts and scaling. Experimental results show that the average repeatability rate is improved from 46% to approximately 78% when all the enhancements are applied. We also compare our algorithm with other region detectors on a set of sequences of real images, and our detector outperforms most of the state of the art detectors
    corecore