107 research outputs found

    EchoGNN: Explainable Ejection Fraction Estimation with Graph Neural Networks

    Full text link
    Ejection fraction (EF) is a key indicator of cardiac function, allowing identification of patients prone to heart dysfunctions such as heart failure. EF is estimated from cardiac ultrasound videos known as echocardiograms (echo) by manually tracing the left ventricle and estimating its volume on certain frames. These estimations exhibit high inter-observer variability due to the manual process and varying video quality. Such sources of inaccuracy and the need for rapid assessment necessitate reliable and explainable machine learning techniques. In this work, we introduce EchoGNN, a model based on graph neural networks (GNNs) to estimate EF from echo videos. Our model first infers a latent echo-graph from the frames of one or multiple echo cine series. It then estimates weights over nodes and edges of this graph, indicating the importance of individual frames that aid EF estimation. A GNN regressor uses this weighted graph to predict EF. We show, qualitatively and quantitatively, that the learned graph weights provide explainability through identification of critical frames for EF estimation, which can be used to determine when human intervention is required. On EchoNet-Dynamic public EF dataset, EchoGNN achieves EF prediction performance that is on par with state of the art and provides explainability, which is crucial given the high inter-observer variability inherent in this task.Comment: Published in MICCAI 202

    Manifold DivideMix: A Semi-Supervised Contrastive Learning Framework for Severe Label Noise

    Full text link
    Deep neural networks have proven to be highly effective when large amounts of data with clean labels are available. However, their performance degrades when training data contains noisy labels, leading to poor generalization on the test set. Real-world datasets contain noisy label samples that either have similar visual semantics to other classes (in-distribution) or have no semantic relevance to any class (out-of-distribution) in the dataset. Most state-of-the-art methods leverage ID labeled noisy samples as unlabeled data for semi-supervised learning, but OOD labeled noisy samples cannot be used in this way because they do not belong to any class within the dataset. Hence, in this paper, we propose incorporating the information from all the training data by leveraging the benefits of self-supervised training. Our method aims to extract a meaningful and generalizable embedding space for each sample regardless of its label. Then, we employ a simple yet effective K-nearest neighbor method to remove portions of out-of-distribution samples. By discarding these samples, we propose an iterative "Manifold DivideMix" algorithm to find clean and noisy samples, and train our model in a semi-supervised way. In addition, we propose "MixEMatch", a new algorithm for the semi-supervised step that involves mixup augmentation at the input and final hidden representations of the model. This will extract better representations by interpolating both in the input and manifold spaces. Extensive experiments on multiple synthetic-noise image benchmarks and real-world web-crawled datasets demonstrate the effectiveness of our proposed framework. Code is available at https://github.com/Fahim-F/ManifoldDivideMix

    Fusion analysis of functional MRI data for classification of individuals based on patterns of activation.

    Get PDF
    Classification of individuals based on patterns of brain activity observed in functional MRI contrasts may be helpful for diagnosis of neurological disorders. Prior work for classification based on these patterns have primarily focused on using a single contrast, which does not take advantage of complementary information that may be available in multiple contrasts. Where multiple contrasts are used, the objective has been only to identify the joint, distinct brain activity patterns that differ between groups of subjects; not to use the information to classify individuals. Here, we use joint Independent Component Analysis (jICA) within a Support Vector Machine (SVM) classification method, and take advantage of the relative contribution of activation patterns generated from multiple fMRI contrasts to improve classification accuracy. Young (age: 19-26) and older (age: 57-73) adults (16 each) were scanned while listening to noise alone and to speech degraded with noise, half of which contained meaningful context that could be used to enhance intelligibility. Functional contrasts based on these conditions (and a silent baseline condition) were used within jICA to generate spatially independent joint activation sources and their corresponding modulation profiles. Modulation profiles were used within a non-linear SVM framework to classify individuals as young or older. Results demonstrate that a combination of activation maps across the multiple contrasts yielded an area under ROC curve of 0.86, superior to classification resulting from individual contrasts. Moreover, class separability, measured by a divergence criterion, was substantially higher when using the combination of activation maps
    • …
    corecore