985 research outputs found
Classification of Infant Sleep/Wake States: Cross-Attention among Large Scale Pretrained Transformer Networks using Audio, ECG, and IMU Data
Infant sleep is critical to brain and behavioral development. Prior studies
on infant sleep/wake classification have been largely limited to reliance on
expensive and burdensome polysomnography (PSG) tests in the laboratory or
wearable devices that collect single-modality data. To facilitate data
collection and accuracy of detection, we aimed to advance this field of study
by using a multi-modal wearable device, LittleBeats (LB), to collect audio,
electrocardiogram (ECG), and inertial measurement unit (IMU) data among a
cohort of 28 infants. We employed a 3-branch (audio/ECG/IMU) large scale
transformer-based neural network (NN) to demonstrate the potential of such
multi-modal data. We pretrained each branch independently with its respective
modality, then finetuned the model by fusing the pretrained transformer layers
with cross-attention. We show that multi-modal data significantly improves
sleep/wake classification (accuracy = 0.880), compared with use of a single
modality (accuracy = 0.732). Our approach to multi-modal mid-level fusion may
be adaptable to a diverse range of architectures and tasks, expanding future
directions of infant behavioral research.Comment: Preprint for APSIPA202
Recommended from our members
Digital phenotyping through multimodal, unobtrusive sensing
The growing adoption of multimodal wearable and mobile devices, such as smartphones and wrist-worn watches has generated an increase in the collection of physiological and behavioural data at scale. This digital phenotyping data enables researchers to make inferences regarding users’ physical and mental health at scale, for the first time. However, translating this data into actionable insights requires computational approaches that turn unlabelled, multimodal time-series sensor data into validated measures that can be interpreted at scale.
This thesis describes the derivation of novel computational methods that leverage digital phenotyping data from wearable devices in large-scale populations to infer physical behaviours. These methods combine insights from signal processing, data mining and machine learning alongside domain knowledge in physical activity and sleep epidemiology. First, the inference of sleeping windows in free-living conditions through a heart rate sensing approach is explored. This algorithm is particularly valuable in the absence of ground truth or sleep diaries given its simplicity, adaptability and capacity for personalization. I then explore multistage sleep classification through combined movement and cardiac wearable sensing and machine learning. Further, I demonstrate that postural changes detected through wrist accelerometers can inform habitual behaviours and are valuable complements to traditional, intensity-based physical activity metrics. I then leverage the concomitant responses of heart rate to physical activity that can be captured through multimodal wearable sensors through a self-supervised training task. The resulting embeddings from this task are shown to be useful for the downstream classification of demographic factors, BMI, energy expenditure and cardiorespiratory fitness. Finally, I describe a deep learning model for the adaptive inference of cardiorespiratory fitness (VO2max) using wearable data in free living conditions. I demonstrate the robustness of the model in a large UK population and show the models’ adaptability by evaluating its performance in a subset of the population with repeated measures ~6 years after the original recordings.
Together, this work increases the potential of multimodal wearable and mobile sensors for physical activity and behavioural inferences in population studies. In particular, this thesis showcases the potential of using wearable devices to make valuable physical activity, sleep and fitness inferences in large cohort studies. Given the nature of the data collected and the fact that most of this data is currently generated by commercial providers and not research institutes, laying the foundations for responsible data governance and ethical use of these technologies will be critical to building trust and enabling the development of the field of digital phenotyping.I was funded by GlaxoSmithKline and the Engineering and Physical Sciences Research Council. I was also supported by the Alan Turing Institute through their Enrichment Scheme
Towards More Accurate Automatic Sleep Staging via Deep Transfer Learning.
BACKGROUND: Despite recent significant progress in the development of automatic sleep staging methods, building a good model still remains a big challenge for sleep studies with a small cohort due to the data-variability and data-inefficiency issues. This work presents a deep transfer learning approach to overcome these issues and enable transferring knowledge from a large dataset to a small cohort for automatic sleep staging. METHODS: We start from a generic end-to-end deep learning framework for sequence-to-sequence sleep staging and derive two networks as the means for transfer learning. The networks are first trained in the source domain (i.e. the large database). The pretrained networks are then finetuned in the target domain (i.e. the small cohort) to complete knowledge transfer. We employ the Montreal Archive of Sleep Studies (MASS) database consisting of 200 subjects as the source domain and study deep transfer learning on three different target domains: the Sleep Cassette subset and the Sleep Telemetry subset of the Sleep-EDF Expanded database, and the Surrey-cEEGrid database. The target domains are purposely adopted to cover different degrees of data mismatch to the source domains. RESULTS: Our experimental results show significant performance improvement on automatic sleep staging on the target domains achieved with the proposed deep transfer learning approach. CONCLUSIONS: These results suggest the efficacy of the proposed approach in addressing the above-mentioned data-variability and data-inefficiency issues. SIGNIFICANCE: As a consequence, it would enable one to improve the quality of automatic sleep staging models when the amount of data is relatively small
MetaSleepLearner: A Pilot Study on Fast Adaptation of Bio-signals-Based Sleep Stage Classifier to New Individual Subject Using Meta-Learning.
Identifying bio-signals based-sleep stages requires time-consuming and tedious labor of skilled clinicians. Deep learning approaches have been introduced in order to challenge the automatic sleep stage classification conundrum. However, the difficulties can be posed in replacing the clinicians with the automatic system due to the differences in many aspects found in individual bio-signals, causing the inconsistency in the performance of the model on every incoming individual. Thus, we aim to explore the feasibility of using a novel approach, capable of assisting the clinicians and lessening the workload. We propose the transfer learning framework, entitled MetaSleepLearner, based on Model Agnostic Meta-Learning (MAML), in order to transfer the acquired sleep staging knowledge from a large dataset to new individual subjects. The framework was demonstrated to require the labelling of only a few sleep epochs by the clinicians and allow the remainder to be handled by the system. Layer-wise Relevance Propagation (LRP) was also applied to understand the learning course of our approach. In all acquired datasets, in comparison to the conventional approach, MetaSleepLearner achieved a range of 5.4% to 17.7% improvement with statistical difference in the mean of both approaches. The illustration of the model interpretation after the adaptation to each subject also confirmed that the performance was directed towards reasonable learning. MetaSleepLearner outperformed the conventional approaches as a result from the fine-tuning using the recordings of both healthy subjects and patients. This is the first work that investigated a non-conventional pre-training method, MAML, resulting in a possibility for human-machine collaboration in sleep stage classification and easing the burden of the clinicians in labelling the sleep stages through only several epochs rather than an entire recording
A Knowledge Distillation Framework For Enhancing Ear-EEG Based Sleep Staging With Scalp-EEG Data
Sleep plays a crucial role in the well-being of human lives. Traditional
sleep studies using Polysomnography are associated with discomfort and often
lower sleep quality caused by the acquisition setup. Previous works have
focused on developing less obtrusive methods to conduct high-quality sleep
studies, and ear-EEG is among popular alternatives. However, the performance of
sleep staging based on ear-EEG is still inferior to scalp-EEG based sleep
staging. In order to address the performance gap between scalp-EEG and ear-EEG
based sleep staging, we propose a cross-modal knowledge distillation strategy,
which is a domain adaptation approach. Our experiments and analysis validate
the effectiveness of the proposed approach with existing architectures, where
it enhances the accuracy of the ear-EEG based sleep staging by 3.46% and
Cohen's kappa coefficient by a margin of 0.038.Comment: Code available at :
https://github.com/Mithunjha/EarEEG_KnowledgeDistillatio
An explainable deep-learning architecture for pediatric sleep apnea identification from overnight airflow and oximetry signals
Producción CientíficaDeep-learning algorithms have been proposed to analyze overnight airflow (AF) and oximetry (SpO2) signals to simplify the diagnosis of pediatric obstructive sleep apnea (OSA), but current algorithms are hardly interpretable. Explainable artificial intelligence (XAI) algorithms can clarify the models-derived predictions on these signals, enhancing their diagnostic trustworthiness. Here, we assess an explainable architecture that combines convolutional and recurrent neural networks (CNN + RNN) to detect pediatric OSA and its severity. AF and SpO2 were obtained from the Childhood Adenotonsillectomy Trial (CHAT) public database (n = 1,638) and a proprietary database (n = 974). These signals were arranged in 30-min segments and processed by the CNN + RNN architecture to derive the number of apneic events per segment. The apnea-hypopnea index (AHI) was computed from the CNN + RNN-derived estimates and grouped into four OSA severity levels. The Gradient-weighted Class Activation Mapping (Grad-CAM) XAI algorithm was used to identify and interpret novel OSA-related patterns of interest. The AHI regression reached very high agreement (intraclass correlation coefficient > 0.9), while OSA severity classification achieved 4-class accuracies 74.51% and 62.31%, and 4-class Cohen’s Kappa 0.6231 and 0.4495, in CHAT and the private datasets, respectively. All diagnostic accuracies on increasing AHI cutoffs (1, 5 and 10 events/h) surpassed 84%. The Grad-CAM heatmaps revealed that the model focuses on sudden AF cessations and SpO2 drops to detect apneas and hypopneas with desaturations, and often discards patterns of hypopneas linked to arousals. Therefore, an interpretable CNN + RNN model to analyze AF and SpO2 can be helpful as a diagnostic alternative in symptomatic children at risk of OSA.Ministerio de Ciencia e Innovación /AEI/10.13039/501100011033/ FEDER (grants PID2020-115468RB-I00 and PDC2021-120775-I00)CIBER -Consorcio Centro de Investigación Biomédica en Red- (CB19/01/00012), Instituto de Salud Carlos IIINational Institutes of Health (HL083075, HL083129, UL1-RR-024134, UL1 RR024989)National Heart, Lung, and Blood Institute (R24 HL114473, 75N92019R002)Ministerio de Ciencia e Innovación - Agencia Estatal de Investigación- “Ramón y Cajal” grant (RYC2019-028566-I
Data-efficient Deep Learning Approach for Single-Channel EEG-Based Sleep Stage Classification with Model Interpretability
Sleep, a fundamental physiological process, occupies a significant portion of
our lives. Accurate classification of sleep stages serves as a crucial tool for
evaluating sleep quality and identifying probable sleep disorders. Our work
introduces a novel methodology that utilizes a SE-Resnet-Bi-LSTM architecture
to classify sleep into five separate stages. The classification process is
based on the analysis of single-channel electroencephalograms (EEGs). The
suggested framework consists of two fundamental elements: a feature extractor
that utilizes SE-ResNet, and a temporal context encoder that uses stacks of
Bi-LSTM units. The effectiveness of our approach is substantiated by thorough
assessments conducted on three different datasets, namely SleepEDF-20,
SleepEDF-78, and SHHS. The proposed methodology achieves significant model
performance, with Macro-F1 scores of 82.5, 78.9, and 81.9 for the respective
datasets. We employ 1D-GradCAM visualization as a methodology to elucidate the
decision-making process inherent in our model in the realm of sleep stage
classification. This visualization method not only provides valuable insights
into the model's classification rationale but also aligns its outcomes with the
annotations made by sleep experts. One notable feature of our research lies in
the incorporation of an efficient training approach, which adeptly upholds the
model's resilience in terms of performance. The experimental evaluations
provide a comprehensive evaluation of the effectiveness of our proposed model
in comparison to the existing approaches, highlighting its potential for
practical applications
- …