182 research outputs found
A Survey on Deep Learning in Medical Image Analysis
Deep learning algorithms, in particular convolutional networks, have rapidly
become a methodology of choice for analyzing medical images. This paper reviews
the major deep learning concepts pertinent to medical image analysis and
summarizes over 300 contributions to the field, most of which appeared in the
last year. We survey the use of deep learning for image classification, object
detection, segmentation, registration, and other tasks and provide concise
overviews of studies per application area. Open challenges and directions for
future research are discussed.Comment: Revised survey includes expanded discussion section and reworked
introductory section on common deep architectures. Added missed papers from
before Feb 1st 201
PadChest: A large chest x-ray image dataset with multi-label annotated reports
We present a labeled large-scale, high resolution chest x-ray dataset for the
automated exploration of medical images along with their associated reports.
This dataset includes more than 160,000 images obtained from 67,000 patients
that were interpreted and reported by radiologists at Hospital San Juan
Hospital (Spain) from 2009 to 2017, covering six different position views and
additional information on image acquisition and patient demography. The reports
were labeled with 174 different radiographic findings, 19 differential
diagnoses and 104 anatomic locations organized as a hierarchical taxonomy and
mapped onto standard Unified Medical Language System (UMLS) terminology. Of
these reports, 27% were manually annotated by trained physicians and the
remaining set was labeled using a supervised method based on a recurrent neural
network with attention mechanisms. The labels generated were then validated in
an independent test set achieving a 0.93 Micro-F1 score. To the best of our
knowledge, this is one of the largest public chest x-ray database suitable for
training supervised models concerning radiographs, and the first to contain
radiographic reports in Spanish. The PadChest dataset can be downloaded from
http://bimcv.cipf.es/bimcv-projects/padchest/
Automated detection, labelling and radiological grading of clinical spinal MRIs
Spinal magnetic resonance (MR) scans are a vital tool for diagnosing the cause of back pain for many diseases and conditions. However, interpreting clinically useful information from these scans can be challenging, time-consuming and hard to reproduce across different radiologists. In this paper, we alleviate these problems by introducing a multi-stage automated pipeline for analysing spinal MR scans. This pipeline first detects and labels vertebral bodies across several commonly used sequences (e.g. T1w, T2w and STIR) and fields of view (e.g. lumbar, cervical, whole spine). Using these detections it then performs automated diagnosis for several spinal disorders, including intervertebral disc degenerative changes in T1w and T2w lumbar scans, and spinal metastases, cord compression and vertebral fractures. To achieve this, we propose a new method of vertebrae detection and labelling, using vector fields to group together detected vertebral landmarks and a language-modelling inspired beam search to determine the corresponding levels of the detections. We also employ a new transformer-based architecture to perform radiological grading which incorporates context from multiple vertebrae and sequences, as a real radiologist would. The performance of each stage of the pipeline is tested in isolation on several clinical datasets, each consisting of 66 to 421 scans. The outputs are compared to manual annotations of expert radiologists, demonstrating accurate vertebrae detection across a range of scan parameters. Similarly, the model’s grading predictions for various types of disc degeneration and detection of spinal metastases closely match those of an expert radiologist. To aid future research, our code and trained models are made publicly available
A Convolutional Approach to Vertebrae Detection and Labelling in Whole Spine MRI
We propose a novel convolutional method for the detection and identification
of vertebrae in whole spine MRIs. This involves using a learnt vector field to
group detected vertebrae corners together into individual vertebral bodies and
convolutional image-to-image translation followed by beam search to label
vertebral levels in a self-consistent manner. The method can be applied without
modification to lumbar, cervical and thoracic-only scans across a range of
different MR sequences. The resulting system achieves 98.1% detection rate and
96.5% identification rate on a challenging clinical dataset of whole spine
scans and matches or exceeds the performance of previous systems on lumbar-only
scans. Finally, we demonstrate the clinical applicability of this method, using
it for automated scoliosis detection in both lumbar and whole spine MR scans.Comment: Accepted full paper to Medical Image Computing and Computer Assisted
Intervention 2020. 11 pages plus appendi
SenseCare: A Research Platform for Medical Image Informatics and Interactive 3D Visualization
Clinical research on smart healthcare has an increasing demand for
intelligent and clinic-oriented medical image computing algorithms and
platforms that support various applications. To this end, we have developed
SenseCare research platform for smart healthcare, which is designed to boost
translational research on intelligent diagnosis and treatment planning in
various clinical scenarios. To facilitate clinical research with Artificial
Intelligence (AI), SenseCare provides a range of AI toolkits for different
tasks, including image segmentation, registration, lesion and landmark
detection from various image modalities ranging from radiology to pathology. In
addition, SenseCare is clinic-oriented and supports a wide range of clinical
applications such as diagnosis and surgical planning for lung cancer, pelvic
tumor, coronary artery disease, etc. SenseCare provides several appealing
functions and features such as advanced 3D visualization, concurrent and
efficient web-based access, fast data synchronization and high data security,
multi-center deployment, support for collaborative research, etc. In this
paper, we will present an overview of SenseCare as an efficient platform
providing comprehensive toolkits and high extensibility for intelligent image
analysis and clinical research in different application scenarios.Comment: 11 pages, 10 figure
Anatomy-Aware Inference of the 3D Standing Spine Posture from 2D Radiographs
An important factor for the development of spinal degeneration, pain and the outcome of spinal surgery is known to be the balance of the spine. It must be analyzed in an upright, standing position to ensure physiological loading conditions and visualize load-dependent deformations. Despite the complex 3D shape of the spine, this analysis is currently performed using 2D radiographs, as all frequently used 3D imaging techniques require the patient to be scanned in a prone position. To overcome this limitation, we propose a deep neural network to reconstruct the 3D spinal pose in an upright standing position, loaded naturally. Specifically, we propose a novel neural network architecture, which takes orthogonal 2D radiographs and infers the spine’s 3D posture using vertebral shape priors. In this work, we define vertebral shape priors using an atlas and a spine shape prior, incorporating both into our proposed network architecture. We validate our architecture on digitally reconstructed radiographs, achieving a 3D reconstruction Dice of 0.95, indicating an almost perfect 2D-to-3D domain translation. Validating the reconstruction accuracy of a 3D standing spine on real data is infeasible due to the lack of a valid ground truth. Hence, we design a novel experiment for this purpose, using an orientation invariant distance metric, to evaluate our model’s ability to synthesize full-3D, upright, and patient-specific spine models. We compare the synthesized spine shapes from clinical upright standing radiographs to the same patient’s 3D spinal posture in the prone position from CT
Lumbar spine segmentation in MR images: a dataset and a public benchmark
This paper presents a large publicly available multi-center lumbar spine
magnetic resonance imaging (MRI) dataset with reference segmentations of
vertebrae, intervertebral discs (IVDs), and spinal canal. The dataset includes
447 sagittal T1 and T2 MRI series from 218 patients with a history of low back
pain. It was collected from four different hospitals and was divided into a
training (179 patients) and validation (39 patients) set. An iterative data
annotation approach was used by training a segmentation algorithm on a small
part of the dataset, enabling semi-automatic segmentation of the remaining
images. The algorithm provided an initial segmentation, which was subsequently
reviewed, manually corrected, and added to the training data. We provide
reference performance values for this baseline algorithm and nnU-Net, which
performed comparably. We set up a continuous segmentation challenge to allow
for a fair comparison of different segmentation algorithms. This study may
encourage wider collaboration in the field of spine segmentation, and improve
the diagnostic value of lumbar spine MRI
- …