7,943 research outputs found
Joint analysis of functional genomic data and genome-wide association studies of 18 human traits
Annotations of gene structures and regulatory elements can inform genome-wide
association studies (GWAS). However, choosing the relevant annotations for
interpreting an association study of a given trait remains challenging. We
describe a statistical model that uses association statistics computed across
the genome to identify classes of genomic element that are enriched or depleted
for loci that influence a trait. The model naturally incorporates multiple
types of annotations. We applied the model to GWAS of 18 human traits,
including red blood cell traits, platelet traits, glucose levels, lipid levels,
height, BMI, and Crohn's disease. For each trait, we evaluated the relevance of
450 different genomic annotations, including protein-coding genes, enhancers,
and DNase-I hypersensitive sites in over a hundred tissues and cell lines. We
show that the fraction of phenotype-associated SNPs that influence protein
sequence ranges from around 2% (for platelet volume) up to around 20% (for LDL
cholesterol); that repressed chromatin is significantly depleted for SNPs
associated with several traits; and that cell type-specific DNase-I
hypersensitive sites are enriched for SNPs associated with several traits (for
example, the spleen in platelet volume). Finally, by re-weighting each GWAS
using information from functional genomics, we increase the number of loci with
high-confidence associations by around 5%.Comment: Fixed typos, included minor clarification
A retrospective segmentation analysis of placental volume by magnetic resonance imaging from first trimester to term gestation
Background
Abnormalities of the placenta affect 5–7% of pregnancies. Because disturbances in fetal growth are often preceded by dysfunction of the placenta or attenuation of its normal expansion, placental health warrants careful surveillance. There are limited normative data available for placental volume by MRI.
Objective
To determine normative ranges of placental volume by MRI throughout gestation.
Materials and methods
In this cross-sectional retrospective analysis, we reviewed MRI examinations of pregnant females obtained between 2002 and 2017 at a single institution. We performed semi-automated segmentation of the placenta in images obtained in patients with no radiologic evidence of maternal or fetal pathology, using the Philips Intellispace Tumor Tracking Tool.
Results
Placental segmentation was performed in 112 women and had a high degree of interrater reliability (single-measure intraclass correlation coefficient =0.978 with 95% confidence interval [CI] 0.956, 0.989; P<0.001). Normative data on placental volume by MRI increased nonlinearly from 6 weeks to 39 weeks of gestation, with wider variability of placental volume at higher gestational age (GA). We fit placental volumetric data to a polynomial curve of third order described as placental volume = –0.02*GA3 + 1.6*GA2 – 13.3*GA + 8.3. Placental volume showed positive correlation with estimated fetal weight (P=0.03) and birth weight (P=0.05).
Conclusion
This study provides normative placental volume by MRI from early first trimester to term gestation. Deviations in placental volume from normal might prove to be an imaging biomarker of adverse fetal health and neonatal outcome, and further studies are needed to more fully understand this metric. Assessment of placental volume should be considered in all routine fetal MRI examinations
Deep Learning in Cardiology
The medical field is creating large amount of data that physicians are unable
to decipher and use efficiently. Moreover, rule-based expert systems are
inefficient in solving complicated medical tasks or for creating insights using
big data. Deep learning has emerged as a more accurate and effective technology
in a wide range of medical problems such as diagnosis, prediction and
intervention. Deep learning is a representation learning method that consists
of layers that transform the data non-linearly, thus, revealing hierarchical
relationships and structures. In this review we survey deep learning
application papers that use structured data, signal and imaging modalities from
cardiology. We discuss the advantages and limitations of applying deep learning
in cardiology that also apply in medicine in general, while proposing certain
directions as the most viable for clinical use.Comment: 27 pages, 2 figures, 10 table
SonoNet: Real-Time Detection and Localisation of Fetal Standard Scan Planes in Freehand Ultrasound
Identifying and interpreting fetal standard scan planes during 2D ultrasound
mid-pregnancy examinations are highly complex tasks which require years of
training. Apart from guiding the probe to the correct location, it can be
equally difficult for a non-expert to identify relevant structures within the
image. Automatic image processing can provide tools to help experienced as well
as inexperienced operators with these tasks. In this paper, we propose a novel
method based on convolutional neural networks which can automatically detect 13
fetal standard views in freehand 2D ultrasound data as well as provide a
localisation of the fetal structures via a bounding box. An important
contribution is that the network learns to localise the target anatomy using
weak supervision based on image-level labels only. The network architecture is
designed to operate in real-time while providing optimal output for the
localisation task. We present results for real-time annotation, retrospective
frame retrieval from saved videos, and localisation on a very large and
challenging dataset consisting of images and video recordings of full clinical
anomaly screenings. We found that the proposed method achieved an average
F1-score of 0.798 in a realistic classification experiment modelling real-time
detection, and obtained a 90.09% accuracy for retrospective frame retrieval.
Moreover, an accuracy of 77.8% was achieved on the localisation task.Comment: 12 pages, 8 figures, published in IEEE Transactions in Medical
Imagin
Attention Gated Networks: Learning to Leverage Salient Regions in Medical Images
We propose a novel attention gate (AG) model for medical image analysis that
automatically learns to focus on target structures of varying shapes and sizes.
Models trained with AGs implicitly learn to suppress irrelevant regions in an
input image while highlighting salient features useful for a specific task.
This enables us to eliminate the necessity of using explicit external
tissue/organ localisation modules when using convolutional neural networks
(CNNs). AGs can be easily integrated into standard CNN models such as VGG or
U-Net architectures with minimal computational overhead while increasing the
model sensitivity and prediction accuracy. The proposed AG models are evaluated
on a variety of tasks, including medical image classification and segmentation.
For classification, we demonstrate the use case of AGs in scan plane detection
for fetal ultrasound screening. We show that the proposed attention mechanism
can provide efficient object localisation while improving the overall
prediction performance by reducing false positives. For segmentation, the
proposed architecture is evaluated on two large 3D CT abdominal datasets with
manual annotations for multiple organs. Experimental results show that AG
models consistently improve the prediction performance of the base
architectures across different datasets and training sizes while preserving
computational efficiency. Moreover, AGs guide the model activations to be
focused around salient regions, which provides better insights into how model
predictions are made. The source code for the proposed AG models is publicly
available.Comment: Accepted for Medical Image Analysis (Special Issue on Medical Imaging
with Deep Learning). arXiv admin note: substantial text overlap with
arXiv:1804.03999, arXiv:1804.0533
A Survey on Deep Learning in Medical Image Analysis
Deep learning algorithms, in particular convolutional networks, have rapidly
become a methodology of choice for analyzing medical images. This paper reviews
the major deep learning concepts pertinent to medical image analysis and
summarizes over 300 contributions to the field, most of which appeared in the
last year. We survey the use of deep learning for image classification, object
detection, segmentation, registration, and other tasks and provide concise
overviews of studies per application area. Open challenges and directions for
future research are discussed.Comment: Revised survey includes expanded discussion section and reworked
introductory section on common deep architectures. Added missed papers from
before Feb 1st 201
Temporal HeartNet: Towards Human-Level Automatic Analysis of Fetal Cardiac Screening Video
We present an automatic method to describe clinically useful information
about scanning, and to guide image interpretation in ultrasound (US) videos of
the fetal heart. Our method is able to jointly predict the visibility, viewing
plane, location and orientation of the fetal heart at the frame level. The
contributions of the paper are three-fold: (i) a convolutional neural network
architecture is developed for a multi-task prediction, which is computed by
sliding a 3x3 window spatially through convolutional maps. (ii) an anchor
mechanism and Intersection over Union (IoU) loss are applied for improving
localization accuracy. (iii) a recurrent architecture is designed to
recursively compute regional convolutional features temporally over sequential
frames, allowing each prediction to be conditioned on the whole video. This
results in a spatial-temporal model that precisely describes detailed heart
parameters in challenging US videos. We report results on a real-world clinical
dataset, where our method achieves performance on par with expert annotations.Comment: To appear in MICCAI, 201
Explainable Anatomical Shape Analysis through Deep Hierarchical Generative Models
Quantification of anatomical shape changes currently relies on scalar global indexes which are largely insensitive to regional or asymmetric modifications. Accurate assessment of pathology-driven anatomical remodeling is a crucial step for the diagnosis and treatment of many conditions. Deep learning approaches have recently achieved wide success in the analysis of medical images, but they lack interpretability in the feature extraction and decision processes. In this work, we propose a new interpretable deep learning model for shape analysis. In particular, we exploit deep generative networks to model a population of anatomical segmentations through a hierarchy of conditional latent variables. At the highest level of this hierarchy, a two-dimensional latent space is simultaneously optimised to discriminate distinct clinical conditions, enabling the direct visualisation of the classification space. Moreover, the anatomical variability encoded by this discriminative latent space can be visualised in the segmentation space thanks to the generative properties of the model, making the classification task transparent. This approach yielded high accuracy in the categorisation of healthy and remodelled left ventricles when tested on unseen segmentations from our own multi-centre dataset as well as in an external validation set, and on hippocampi from healthy controls and patients with Alzheimer's disease when tested on ADNI data. More importantly, it enabled the visualisation in three-dimensions of both global and regional anatomical features which better discriminate between the conditions under exam. The proposed approach scales effectively to large populations, facilitating high-throughput analysis of normal anatomy and pathology in large-scale studies of volumetric imaging
- …