315 research outputs found
Deep Convolutional Neural Networks for Interpretable Analysis of EEG Sleep Stage Scoring
Sleep studies are important for diagnosing sleep disorders such as insomnia,
narcolepsy or sleep apnea. They rely on manual scoring of sleep stages from raw
polisomnography signals, which is a tedious visual task requiring the workload
of highly trained professionals. Consequently, research efforts to purse for an
automatic stage scoring based on machine learning techniques have been carried
out over the last years. In this work, we resort to multitaper spectral
analysis to create visually interpretable images of sleep patterns from EEG
signals as inputs to a deep convolutional network trained to solve visual
recognition tasks. As a working example of transfer learning, a system able to
accurately classify sleep stages in new unseen patients is presented.
Evaluations in a widely-used publicly available dataset favourably compare to
state-of-the-art results, while providing a framework for visual interpretation
of outcomes.Comment: 8 pages, 1 figure, 2 tables, IEEE 2017 International Workshop on
Machine Learning for Signal Processin
Deep residual networks for automatic sleep stage classification of raw polysomnographic waveforms
We have developed an automatic sleep stage classification algorithm based on
deep residual neural networks and raw polysomnogram signals. Briefly, the raw
data is passed through 50 convolutional layers before subsequent classification
into one of five sleep stages. Three model configurations were trained on 1850
polysomnogram recordings and subsequently tested on 230 independent recordings.
Our best performing model yielded an accuracy of 84.1% and a Cohen's kappa of
0.746, improving on previous reported results by other groups also using only
raw polysomnogram data. Most errors were made on non-REM stage 1 and 3
decisions, errors likely resulting from the definition of these stages. Further
testing on independent cohorts is needed to verify performance for clinical
use
Automatic sleep staging of EEG signals: recent development, challenges, and future directions.
Modern deep learning holds a great potential to transform clinical studies of human sleep. Teaching a machine to carry out routine tasks would be a tremendous reduction in workload for clinicians. Sleep staging, a fundamental step in sleep practice, is a suitable task for this and will be the focus in this article. Recently, automatic sleep-staging systems have been trained to mimic manual scoring, leading to similar performance to human sleep experts, at least on scoring of healthy subjects. Despite tremendous progress, we have not seen automatic sleep scoring adopted widely in clinical environments. This review aims to provide the shared view of the authors on the most recent state-of-the-art developments in automatic sleep staging, the challenges that still need to be addressed, and the future directions needed for automatic sleep scoring to achieve clinical value
STQS:Interpretable multi-modal Spatial-Temporal-seQuential model for automatic Sleep scoring
Sleep scoring is an important step for the detection of sleep disorders and usually performed by visual analysis. Since manual sleep scoring is time consuming, machine-learning based approaches have been proposed. Though efficient, these algorithms are black-box in nature and difficult to interpret by clinicians. In this paper, we propose a deep learning architecture for multi-modal sleep scoring, investigate the model's decision making process, and compare the model's reasoning with the annotation guidelines in the AASM manual. Our architecture, called STQS, uses convolutional neural networks (CNN) to automatically extract spatio-temporal features from 3 modalities (EEG, EOG and EMG), a bidirectional long short-term memory (Bi-LSTM) to extract sequential information, and residual connections to combine spatio-temporal and sequential features. We evaluated our model on two large datasets, obtaining an accuracy of 85% and 77% and a macro F1 score of 79% and 73% on SHHS and an in-house dataset, respectively. We further quantify the contribution of various architectural components and conclude that adding LSTM layers improves performance over a spatio-temporal CNN, while adding residual connections does not. Our interpretability results show that the output of the model is well aligned with AASM guidelines, and therefore, the model's decisions correspond to domain knowledge. We also compare multi-modal models and single-channel models and suggest that future research should focus on improving multi-modal models
Towards a Deeper Understanding of Sleep Stages through their Representation in the Latent Space of Variational Autoencoders
Artificial neural networks show great success in sleep stage classification, with an accuracy comparable to human scoring. While their ability to learn from labelled electroencephalography (EEG) signals is widely researched, the underlying learning processes remain unexplored. Variational autoencoders can capture the underlying meaning of data by encoding it into a low-dimensional space. Regularizing this space furthermore enables the generation of realistic representations of data from latent space samples. We aimed to show that this model is able to generate realistic sleep EEG. In addition, the generated sequences from different areas of the latent space are shown to have inherent meaning. The current results show the potential of variational autoencoders in understanding sleep EEG data from the perspective of unsupervised machine learning
RED: Deep Recurrent Neural Networks for Sleep EEG Event Detection
The brain electrical activity presents several short events during sleep that
can be observed as distinctive micro-structures in the electroencephalogram
(EEG), such as sleep spindles and K-complexes. These events have been
associated with biological processes and neurological disorders, making them a
research topic in sleep medicine. However, manual detection limits their study
because it is time-consuming and affected by significant inter-expert
variability, motivating automatic approaches. We propose a deep learning
approach based on convolutional and recurrent neural networks for sleep EEG
event detection called Recurrent Event Detector (RED). RED uses one of two
input representations: a) the time-domain EEG signal, or b) a complex
spectrogram of the signal obtained with the Continuous Wavelet Transform (CWT).
Unlike previous approaches, a fixed time window is avoided and temporal context
is integrated to better emulate the visual criteria of experts. When evaluated
on the MASS dataset, our detectors outperform the state of the art in both
sleep spindle and K-complex detection with a mean F1-score of at least 80.9%
and 82.6%, respectively. Although the CWT-domain model obtained a similar
performance than its time-domain counterpart, the former allows in principle a
more interpretable input representation due to the use of a spectrogram. The
proposed approach is event-agnostic and can be used directly to detect other
types of sleep events.Comment: 8 pages, 5 figures. In proceedings of the 2020 International Joint
Conference on Neural Networks (IJCNN 2020
- âŠ