119 research outputs found
Evaluation of machine learning algorithms for Health and Wellness applications: a tutorial
Research on decision support applications in healthcare, such as those
related to diagnosis, prediction, treatment planning, etc., have seen
enormously increased interest recently. This development is thanks to the
increase in data availability as well as advances in artificial intelligence
and machine learning research. Highly promising research examples are published
daily. However, at the same time, there are some unrealistic expectations with
regards to the requirements for reliable development and objective validation
that is needed in healthcare settings. These expectations may lead to unmet
schedules and disappointments (or non-uptake) at the end-user side. It is the
aim of this tutorial to provide practical guidance on how to assess performance
reliably and efficiently and avoid common traps. Instead of giving a list of
do's and don't s, this tutorial tries to build a better understanding behind
these do's and don't s and presents both the most relevant performance
evaluation criteria as well as how to compute them. Along the way, we will
indicate common mistakes and provide references discussing various topics more
in-depth.Comment: To be published in Computers in Biology and Medicin
Adversarial Distortion Learning for Medical Image Denoising
We present a novel adversarial distortion learning (ADL) for denoising two-
and three-dimensional (2D/3D) biomedical image data. The proposed ADL consists
of two auto-encoders: a denoiser and a discriminator. The denoiser removes
noise from input data and the discriminator compares the denoised result to its
noise-free counterpart. This process is repeated until the discriminator cannot
differentiate the denoised data from the reference. Both the denoiser and the
discriminator are built upon a proposed auto-encoder called Efficient-Unet.
Efficient-Unet has a light architecture that uses the residual blocks and a
novel pyramidal approach in the backbone to efficiently extract and re-use
feature maps. During training, the textural information and contrast are
controlled by two novel loss functions. The architecture of Efficient-Unet
allows generalizing the proposed method to any sort of biomedical data. The 2D
version of our network was trained on ImageNet and tested on biomedical
datasets whose distribution is completely different from ImageNet; so, there is
no need for re-training. Experimental results carried out on magnetic resonance
imaging (MRI), dermatoscopy, electron microscopy and X-ray datasets show that
the proposed method achieved the best on each benchmark. Our implementation and
pre-trained models are available at https://github.com/mogvision/ADL
Regularized bagged canonical component analysis for multiclass learning in brain imaging
Alzheimer’s Disease Neuroimaging Initiative (ADNI) is a Group/Institutional Author. Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report.A fundamental problem of supervised learning algorithms for brain imaging applications is that the number of features far exceeds the number of subjects. In this paper, we propose a combined feature selection and extraction approach for multiclass problems. This method starts with a bagging procedure which calculates the sign consistency of the multivariate analysis (MVA) projection matrix feature-wise to determine the relevance of each feature. This relevance measure provides a parsimonious matrix, which is combined with a hypothesis test to automatically determine the number of selected features. Then, a novel MVA regularized with the sign and magnitude consistency of the features is used to generate a reduced set of summary components providing a compact data description. We evaluated the proposed method with two multiclass brain imaging problems: 1) the classification of the elderly subjects in four classes (cognitively normal, stable mild cognitive impairment (MCI), MCI converting to AD in 3 years, and Alzheimer’s disease) based on structural brain imaging data from the ADNI cohort; 2) the classification of children in 3 classes (typically developing, and 2 types of Attention Deficit/Hyperactivity Disorder (ADHD)) based on functional connectivity. Experimental results confirmed that each brain image (defined by 29.852 features in the ADNI database and 61.425 in the ADHD) could be represented with only 30 − 45% of the original features. Furthermore, this information could be redefined into two or three summary components, providing not only a gain of interpretability but also classification rate improvements when compared to state-of-art reference methods.C. Sevilla-Salcedo and V. Gomez-Verdejo's work has been partly funded by the Spanish MINECO grant TEC2014-52289-R and TEC2017-83838-R as well as KERMES, which is a NoE on kernel methods for structured data, funded by the Spanish Ministry of Economy and Competitiveness, TEC2016-81900-REDT ru. Jussi Tohka's work is supported by the Academy of Finland (grant 316258)
Comparison of feature representations in MRI-based MCI-to-AD conversion prediction
Alzheimer's disease (AD) is a progressive neurological disorder in which the death of brain cells causes memory loss and cognitive decline. The identification of at-risk subjects yet showing no dementia symptoms but who will later convert to AD can be crucial for the effective treatment of AD. For this, Magnetic Resonance Imaging (MRI) is expected to play a crucial role. During recent years, several Machine Learning (ML) approaches to AD-conversion prediction have been proposed using different types of MRI features. However, few studies comparing these different feature representations exist, and the existing ones do not allow to make definite conclusions. We evaluated the performance of various types of MRI features for the conversion prediction: voxel-based features extracted based on voxel-based morphometry, hippocampus volumes, volumes of the entorhinal cortex, and a set of regional volumetric, surface area, and cortical thickness measures across the brain. Regional features consistently yielded the best performance over two classifiers (Support Vector Machines and Regularized Logistic Regression), and two datasets studied. However, the performance difference to other features was not statistically significant. There was a consistent trend of age correction improving the classification performance, but the improvement reached statistical significance only rarely.Data collection and sharing for this project was funded by the Alzheimer's Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). ADNI is funded by the National Institute on Aging, the National Institute of Biomedical Imaging and Bioengineering, and through generous contributions from the following: AbbVie, Alzheimer's Association; Alzheimer's Drug Discovery Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb Company; CereSpir, Inc.; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F. Hoffmann-La Roche Ltd and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics.
J. Tohka's work was supported by the Academy of Finland and V. Gómez-Verdejo's work has been partly funded by the Spanish MINECO grant TEC2014-52289R, TEC2016-81900-REDT/AEI and TEC2017-83838-R
Self-Supervised Super-Resolution Approach for Isotropic Reconstruction of 3D Electron Microscopy Images from Anisotropic Acquisition
Three-dimensional electron microscopy (3DEM) is an essential technique to
investigate volumetric tissue ultra-structure. Due to technical limitations and
high imaging costs, samples are often imaged anisotropically, where resolution
in the axial direction () is lower than in the lateral directions .
This anisotropy 3DEM can hamper subsequent analysis and visualization tasks. To
overcome this limitation, we propose a novel deep-learning (DL)-based
self-supervised super-resolution approach that computationally reconstructs
isotropic 3DEM from the anisotropic acquisition. The proposed DL-based
framework is built upon the U-shape architecture incorporating
vision-transformer (ViT) blocks, enabling high-capability learning of local and
global multi-scale image dependencies. To train the tailored network, we employ
a self-supervised approach. Specifically, we generate pairs of anisotropic and
isotropic training datasets from the given anisotropic 3DEM data. By feeding
the given anisotropic 3DEM dataset in the trained network through our proposed
framework, the isotropic 3DEM is obtained. Importantly, this isotropic
reconstruction approach relies solely on the given anisotropic 3DEM dataset and
does not require pairs of co-registered anisotropic and isotropic 3DEM training
datasets. To evaluate the effectiveness of the proposed method, we conducted
experiments using three 3DEM datasets acquired from brain. The experimental
results demonstrated that our proposed framework could successfully reconstruct
isotropic 3DEM from the anisotropic acquisition
No-Clean-Reference Image Super-Resolution: Application to Electron Microscopy
The inability to acquire clean high-resolution (HR) electron microscopy (EM)
images over a large brain tissue volume hampers many neuroscience studies. To
address this challenge, we propose a deep-learning-based image super-resolution
(SR) approach to computationally reconstruct clean HR 3D-EM with a large field
of view (FoV) from noisy low-resolution (LR) acquisition. Our contributions are
I) Investigating training with no-clean references for and
loss functions; II) Introducing a novel network architecture, named EMSR, for
enhancing the resolution of LR EM images while reducing inherent noise; and,
III) Comparing different training strategies including using acquired LR and HR
image pairs, i.e., real pairs with no-clean references contaminated with real
corruptions, the pairs of synthetic LR and acquired HR, as well as acquired LR
and denoised HR pairs. Experiments with nine brain datasets showed that
training with real pairs can produce high-quality super-resolved results,
demonstrating the feasibility of training with non-clean references for both
loss functions. Additionally, comparable results were observed, both visually
and numerically, when employing denoised and noisy references for training.
Moreover, utilizing the network trained with synthetically generated LR images
from HR counterparts proved effective in yielding satisfactory SR results, even
in certain cases, outperforming training with real pairs. The proposed SR
network was compared quantitatively and qualitatively with several established
SR techniques, showcasing either the superiority or competitiveness of the
proposed method in mitigating noise while recovering fine details.Comment: 13 pages, 12 figures, and 2 table
Multi-Objective Genetic Algorithm for Multi-View Feature Selection
Multi-view datasets offer diverse forms of data that can enhance prediction
models by providing complementary information. However, the use of multi-view
data leads to an increase in high-dimensional data, which poses significant
challenges for the prediction models that can lead to poor generalization.
Therefore, relevant feature selection from multi-view datasets is important as
it not only addresses the poor generalization but also enhances the
interpretability of the models. Despite the success of traditional feature
selection methods, they have limitations in leveraging intrinsic information
across modalities, lacking generalizability, and being tailored to specific
classification tasks. We propose a novel genetic algorithm strategy to overcome
these limitations of traditional feature selection methods for multi-view data.
Our proposed approach, called the multi-view multi-objective feature selection
genetic algorithm (MMFS-GA), simultaneously selects the optimal subset of
features within a view and between views under a unified framework. The MMFS-GA
framework demonstrates superior performance and interpretability for feature
selection on multi-view datasets in both binary and multiclass classification
tasks. The results of our evaluations on three benchmark datasets, including
synthetic and real data, show improvement over the best baseline methods. This
work provides a promising solution for multi-view feature selection and opens
up new possibilities for further research in multi-view datasets
- …