563 research outputs found
Unsupervised domain adaptation for medical imaging segmentation with self-ensembling
Recent advances in deep learning methods have come to define the
state-of-the-art for many medical imaging applications, surpassing even human
judgment in several tasks. Those models, however, when trained to reduce the
empirical risk on a single domain, fail to generalize when applied to other
domains, a very common scenario in medical imaging due to the variability of
images and anatomical structures, even across the same imaging modality. In
this work, we extend the method of unsupervised domain adaptation using
self-ensembling for the semantic segmentation task and explore multiple facets
of the method on a small and realistic publicly-available magnetic resonance
(MRI) dataset. Through an extensive evaluation, we show that self-ensembling
can indeed improve the generalization of the models even when using a small
amount of unlabelled data.Comment: 15 pages, 9 figure
Learning with Limited Labeled Data in Biomedical Domain by Disentanglement and Semi-Supervised Learning
In this dissertation, we are interested in improving the generalization of deep neural networks for biomedical data (e.g., electrocardiogram signal, x-ray images, etc). Although deep neural networks have attained state-of-the-art performance and, thus, deployment across a variety of domains, similar performance in the clinical setting remains challenging due to its ineptness to generalize across unseen data (e.g., new patient cohort).
We address this challenge of generalization in the deep neural network from two perspectives: 1) learning disentangled representations from the deep network, and 2) developing efficient semi-supervised learning (SSL) algorithms using the deep network.
In the former, we are interested in designing specific architectures and objective functions to learn representations, where variations in the data are well separated, i.e., disentangled. In the latter, we are interested in designing regularizers that encourage the underlying neural function\u27s behavior toward a common inductive bias to avoid over-fitting the function to small labeled data.
Our end goal is to improve the generalization of the deep network for the diagnostic model in both of these approaches. In disentangled representations, this translates to appropriately learning latent representations from the data, capturing the observed input\u27s underlying explanatory factors in an independent and interpretable way. With data\u27s expository factors well separated, such disentangled latent space can then be useful for a large variety of tasks and domains within data distribution even with a small amount of labeled data, thus improving generalization. In developing efficient semi-supervised algorithms, this translates to utilizing a large volume of the unlabelled dataset to assist the learning from the limited labeled dataset, commonly encountered situation in the biomedical domain.
By drawing ideas from different areas within deep learning like representation learning (e.g., autoencoder), variational inference (e.g., variational autoencoder), Bayesian nonparametric (e.g., beta-Bernoulli process), learning theory (e.g., analytical learning theory), function smoothing (Lipschitz Smoothness), etc., we propose several leaning algorithms to improve generalization in the associated task. We test our algorithms on real-world clinical data and show that our approach yields significant improvement over existing methods. Moreover, we demonstrate the efficacy of the proposed models in the benchmark data and simulated data to understand different aspects of the proposed learning methods.
We conclude by identifying some of the limitations of the proposed methods, areas of further improvement, and broader future directions for the successful adoption of AI models in the clinical environment
Recommended from our members
Explainable improved ensembling for natural language and vision
Ensemble methods are well-known in machine learning for improving prediction
accuracy. However, they do not adequately discriminate among underlying
component models. The measure of how good a model is can sometimes be estimated
from “why” it made a specific prediction. We propose a novel approach
called Stacking With Auxiliary Features (SWAF) that effectively leverages component
models by integrating such relevant information from context to improve
ensembling. Using auxiliary features, our algorithm learns to rely on systems that
not just agree on an output prediction but also the source or origin of that output.
We demonstrate our approach to challenging structured prediction problems
in Natural Language Processing and Vision including Information Extraction, Object
Detection, and Visual Question Answering. We also present a variant of SWAF
for combining systems that do not have training data in an unsupervised ensemble
with systems that do have training data. Our combined approach obtains a new
state-of-the-art, beating our prior performance on Information Extraction.
The state-of-the-art systems on many AI applications are ensembles of deeplearning
models. These models are hard to interpret and can sometimes make odd
mistakes. Explanations make AI systems more transparent and also justify their
predictions. We propose a scalable approach to generate visual explanations for
ensemble methods using the localization maps of the component systems. Crowdsourced
human evaluation on two new metrics indicates that our ensemble’s explanation
significantly qualitatively outperforms individual systems’ explanations.Computer Science
Extensive Analysis on Generation and Consensus Mechanisms of Clustering Ensemble: A Survey
Data analysis plays a prominent role in interpreting various phenomena. Data mining is the process to hypothesize useful knowledge from the extensive data. Based upon the classical statistical prototypes the data can be exploited beyond the storage and management of the data. Cluster analysis a primary investigation with little or no prior knowledge, consists of research and development across a wide variety of communities. Cluster ensembles are melange of individual solutions obtained from different clusterings to produce final quality clustering which is required in wider applications. The method arises in the perspective of increasing robustness, scalability and accuracy. This paper gives a brief overview of the generation methods and consensus functions included in cluster ensemble. The survey is to analyze the various techniques and cluster ensemble methods
- …