179 research outputs found
A Novel Hybrid Ordinal Learning Model with Health Care Application
Ordinal learning (OL) is a type of machine learning models with broad utility
in health care applications such as diagnosis of different grades of a disease
(e.g., mild, modest, severe) and prediction of the speed of disease progression
(e.g., very fast, fast, moderate, slow). This paper aims to tackle a situation
when precisely labeled samples are limited in the training set due to cost or
availability constraints, whereas there could be an abundance of samples with
imprecise labels. We focus on imprecise labels that are intervals, i.e., one
can know that a sample belongs to an interval of labels but cannot know which
unique label it has. This situation is quite common in health care datasets due
to limitations of the diagnostic instrument, sparse clinical visits, or/and
patient dropout. Limited research has been done to develop OL models with
imprecise/interval labels. We propose a new Hybrid Ordinal Learner (HOL) to
integrate samples with both precise and interval labels to train a robust OL
model. We also develop a tractable and efficient optimization algorithm to
solve the HOL formulation. We compare HOL with several recently developed OL
methods on four benchmarking datasets, which demonstrate the superior
performance of HOL. Finally, we apply HOL to a real-world dataset for
predicting the speed of progressing to Alzheimer's Disease (AD) for individuals
with Mild Cognitive Impairment (MCI) based on a combination of multi-modality
neuroimaging and demographic/clinical datasets. HOL achieves high accuracy in
the prediction and outperforms existing methods. The capability of accurately
predicting the speed of progression to AD for each individual with MCI has the
potential for helping facilitate more individually-optimized interventional
strategies.Comment: 16 pages, 3 figures, 2 table
Machine Learning for Multiclass Classification and Prediction of Alzheimer\u27s Disease
Alzheimer\u27s disease (AD) is an irreversible neurodegenerative disorder and a common form of dementia. This research aims to develop machine learning algorithms that diagnose and predict the progression of AD from multimodal heterogonous biomarkers with a focus placed on the early diagnosis. To meet this goal, several machine learning-based methods with their unique characteristics for feature extraction and automated classification, prediction, and visualization have been developed to discern subtle progression trends and predict the trajectory of disease progression.
The methodology envisioned aims to enhance both the multiclass classification accuracy and prediction outcomes by effectively modeling the interplay between the multimodal biomarkers, handle the missing data challenge, and adequately extract all the relevant features that will be fed into the machine learning framework, all in order to understand the subtle changes that happen in the different stages of the disease. This research will also investigate the notion of multitasking to discover how the two processes of multiclass classification and prediction relate to one another in terms of the features they share and whether they could learn from one another for optimizing multiclass classification and prediction accuracy.
This research work also delves into predicting cognitive scores of specific tests over time, using multimodal longitudinal data. The intent is to augment our prospects for analyzing the interplay between the different multimodal features used in the input space to the predicted cognitive scores. Moreover, the power of modality fusion, kernelization, and tensorization have also been investigated to efficiently extract important features hidden in the lower-dimensional feature space without being distracted by those deemed as irrelevant.
With the adage that a picture is worth a thousand words, this dissertation introduces a unique color-coded visualization system with a fully integrated machine learning model for the enhanced diagnosis and prognosis of Alzheimer\u27s disease. The incentive here is to show that through visualization, the challenges imposed by both the variability and interrelatedness of the multimodal features could be overcome. Ultimately, this form of visualization via machine learning informs on the challenges faced with multiclass classification and adds insight into the decision-making process for a diagnosis and prognosis
Multimodal and multicontrast image fusion via deep generative models
Recently, it has become progressively more evident that classic diagnostic
labels are unable to reliably describe the complexity and variability of
several clinical phenotypes. This is particularly true for a broad range of
neuropsychiatric illnesses (e.g., depression, anxiety disorders, behavioral
phenotypes). Patient heterogeneity can be better described by grouping
individuals into novel categories based on empirically derived sections of
intersecting continua that span across and beyond traditional categorical
borders. In this context, neuroimaging data carry a wealth of spatiotemporally
resolved information about each patient's brain. However, they are usually
heavily collapsed a priori through procedures which are not learned as part of
model training, and consequently not optimized for the downstream prediction
task. This is because every individual participant usually comes with multiple
whole-brain 3D imaging modalities often accompanied by a deep genotypic and
phenotypic characterization, hence posing formidable computational challenges.
In this paper we design a deep learning architecture based on generative models
rooted in a modular approach and separable convolutional blocks to a) fuse
multiple 3D neuroimaging modalities on a voxel-wise level, b) convert them into
informative latent embeddings through heavy dimensionality reduction, c)
maintain good generalizability and minimal information loss. As proof of
concept, we test our architecture on the well characterized Human Connectome
Project database demonstrating that our latent embeddings can be clustered into
easily separable subject strata which, in turn, map to different phenotypical
information which was not included in the embedding creation process. This may
be of aid in predicting disease evolution as well as drug response, hence
supporting mechanistic disease understanding and empowering clinical trials
Investigation of Multi-dimensional Tensor Multi-task Learning for Modeling Alzheimer's Disease Progression
Machine learning (ML) techniques for predicting Alzheimer's disease (AD) progression can significantly assist clinicians and researchers in constructing effective AD prevention and treatment strategies. The main constraints on the performance of current ML approaches are prediction accuracy and stability problems in medical small dataset scenarios, monotonic data formats (loss of multi-dimensional knowledge of the data and loss of correlation knowledge between biomarkers) and biomarker interpretability limitations. This thesis investigates how multi-dimensional information and knowledge from biomarker data integrated with multi-task learning approaches to predict AD progression. Firstly, a novel similarity-based quantification approach is proposed with two components: multi-dimensional knowledge vector construction and amalgamated magnitude-direction quantification of brain structural variation, which considers both the magnitude and directional correlations of structural variation between brain biomarkers and encodes the quantified data as a third-order tensor to address the problem of monotonic data form. Secondly, multi-task learning regression algorithms with the ability to integrate multi-dimensional tensor data and mine MRI data for spatio-temporal structural variation information and knowledge were designed and constructed to improve the accuracy, stability and interpretability of AD progression prediction in medical small dataset scenarios. The algorithm consists of three components: supervised symmetric tensor decomposition for extracting biomarker latent factors, tensor multi-task learning regression and algorithmic regularisation terms. The proposed algorithm aims to extract a set of first-order latent factors from the raw data, each represented by its first biomarker, second biomarker and patient sample dimensions, to elucidate potential factors affecting the variability of the data in an interpretable manner and can be utilised as predictor variables for training the prediction model that regards the prediction of each patient as a task, with each task sharing a set of biomarker latent factors obtained from tensor decomposition. Knowledge sharing between tasks improves the generalisation ability of the model and addresses the problem of sparse medical data. The experimental results demonstrate that the proposed approach achieves superior accuracy and stability in predicting various cognitive scores of AD progression compared to single-task learning, benchmarks and state-of-the-art multi-task regression methods. The proposed approach identifies brain structural variations in patients and the important brain biomarker correlations revealed by the experiments can be utilised as potential indicators for AD early identification
Enhancing Alzheimer Disease Segmentation through Adaptively Regularized Weighted Kernel-Based Clustering
Image segmentation is important in image analysis because it helps to locate objects and boundaries within a picture. This study offers Adaptively Regularized Weighted Kernel-Based Clustering (ARWKC), a unique segmentation technique built exclusively for recovering brain tissue from medical pictures. The proposed approach incorporates adaptive regularization and weighted kernel-based clustering techniques to increase the accuracy and resilience of brain tissue segmentation. The picture is initially preprocessed with the ARWKC method to improve its quality and eliminate any noise or artifacts. The adaptive regularization method is then utilized to effectively deal with the visual variation of brain tissue in clinical images. This adaptive regularization contributes to more accurate and consistent segmentation outcomes. The weighted kernel-based clustering method is then used to find and group pixels with comparable properties, with a focus on brain tissue areas. This clustering approach employs a weighted kernel function that takes into account both geographical closeness and pixel intensities, allowing the algorithm to capture local picture features and improve segmentation accuracy. Extensive experiments were conducted on a collection of medical images to evaluate the efficacy of the ARWKC algorithm. The well-known k-means clustering method, often used in image segmentation applications, was utilized as a benchmark for comparison. In terms of accuracy and resilience for brain tissue segmentation, the experimental findings showed that the ARWKC method surpasses the k-means clustering approach
Alzheimer’s And Parkinson’s Disease Classification Using Deep Learning Based On MRI: A Review
Neurodegenerative disorders present a current challenge for accurate diagnosis and for providing precise prognostic information. Alzheimer’s disease (AD) and Parkinson's disease (PD), may take several years to obtain a definitive diagnosis. Due to the increased aging population in developed countries, neurodegenerative diseases such as AD and PD have become more prevalent and thus new technologies and more accurate tests are needed to improve and accelerate the diagnostic procedure in the early stages of these diseases. Deep learning has shown significant promise in computer-assisted AD and PD diagnosis based on MRI with the widespread use of artificial intelligence in the medical domain. This article analyses and evaluates the effectiveness of existing Deep learning (DL)-based approaches to identify neurological illnesses using MRI data obtained using various modalities, including functional and structural MRI. Several current research issues are identified toward the conclusion, along with several potential future study directions
Multimodal and Multiscale Deep Neural Networks for the Early Diagnosis of Alzheimer's Disease using structural MR and FDG-PET images.
Alzheimer's Disease (AD) is a progressive neurodegenerative disease where biomarkers for disease based on pathophysiology may be able to provide objective measures for disease diagnosis and staging. Neuroimaging scans acquired from MRI and metabolism images obtained by FDG-PET provide in-vivo measurements of structure and function (glucose metabolism) in a living brain. It is hypothesized that combining multiple different image modalities providing complementary information could help improve early diagnosis of AD. In this paper, we propose a novel deep-learning-based framework to discriminate individuals with AD utilizing a multimodal and multiscale deep neural network. Our method delivers 82.4% accuracy in identifying the individuals with mild cognitive impairment (MCI) who will convert to AD at 3 years prior to conversion (86.4% combined accuracy for conversion within 1-3 years), a 94.23% sensitivity in classifying individuals with clinical diagnosis of probable AD, and a 86.3% specificity in classifying non-demented controls improving upon results in published literature
Recommended from our members
From Data to Diagnosis: Efficient prediction of Alzheimer’s disease and related dementias using statistical machine learning
Faced with a rapidly aging population and the rising prevalence of Alzheimer’s disease (AD) and related dementias, the field of dementia research needs to urgently consider screening tools that utilize widely accessible data modalities. We stratified relevant data based upon collection time, expertise needed, cost of administration, and data availability in a variety of clinical settings to form three distinct tiers: low-burden, medium-burden, and high-burden. Low-burden data, for instance, may include demographics, patient history, family history, and simple cognitive tests (e.g., Mini-Mental State Exam), information that could be reasonably obtained by a primary care physician, non-physician practitioner, or caregiver. Medium-burden information may include an in-depth neuropsychological testing battery or structural magnetic resonance imaging scan, both requiring more resources and time to administer. High-burden data would include information not routinely collected in clinical settings due to cost, time, and expertise required, such as the Clinical Dementia Rating. Using methods from statistical machine learning, we investigated disease prediction using low-burden data in three related projects. First, we developed a framework to evaluate whether inexpensive, accessible data can accurately classify AD clinical diagnosis and predict likelihood of progression. Second, we compared different data integration or “fusion” methods to determine optimal strategies for fusing low-burden data with additional neuroimaging data for predicting AD clinical diagnosis. Third, we developed models to infer dementia neuropathology using low-burden data. We found that using low-burden data yielded relatively high clinical diagnosis prediction performance and enabled us to identify subgroups with high- and low-risk of progression within 1.5-3 years. When integrating multimodal data, we observed that using a supervised encoder with intermediate fusion significantly improved performance. For neuropathology prediction, we identified three groups of subjects with significantly different proportions of positive neuropathology lesions, and we discovered that, with at least three clinical visits, using low-burden features yielded comparable prediction performance to using higher cost features. In conclusion, low-burden clinical data can effectively identify at-risk individuals for both AD clinical diagnosis and postmortem dementia neuropathology
Deep Interpretability Methods for Neuroimaging
Brain dynamics are highly complex and yet hold the key to understanding brain function and dysfunction. The dynamics captured by resting-state functional magnetic resonance imaging data are noisy, high-dimensional, and not readily interpretable. The typical approach of reducing this data to low-dimensional features and focusing on the most predictive features comes with strong assumptions and can miss essential aspects of the underlying dynamics. In contrast, introspection of discriminatively trained deep learning models may uncover disorder-relevant elements of the signal at the level of individual time points and spatial locations. Nevertheless, the difficulty of reliable training on high-dimensional but small-sample datasets and the unclear relevance of the resulting predictive markers prevent the widespread use of deep learning in functional neuroimaging. In this dissertation, we address these challenges by proposing a deep learning framework to learn from high-dimensional dynamical data while maintaining stable, ecologically valid interpretations. The developed model is pre-trainable and alleviates the need to collect an enormous amount of neuroimaging samples to achieve optimal training.
We also provide a quantitative validation module, Retain and Retrain (RAR), that can objectively verify the higher predictability of the dynamics learned by the model. Results successfully demonstrate that the proposed framework enables learning the fMRI dynamics directly from small data and capturing compact, stable interpretations of features predictive of function and dysfunction. We also comprehensively reviewed deep interpretability literature in the neuroimaging domain. Our analysis reveals the ongoing trend of interpretability practices in neuroimaging studies and identifies the gaps that should be addressed for effective human-machine collaboration in this domain.
This dissertation also proposed a post hoc interpretability method, Geometrically Guided Integrated Gradients (GGIG), that leverages geometric properties of the functional space as learned by a deep learning model. With extensive experiments and quantitative validation on MNIST and ImageNet datasets, we demonstrate that GGIG outperforms integrated gradients (IG), which is considered to be a popular interpretability method in the literature. As GGIG is able to identify the contours of the discriminative regions in the input space, GGIG may be useful in various medical imaging tasks where fine-grained localization as an explanation is beneficial
- …