3 research outputs found
Rapid annotation of seizures and interictal-ictal-injury continuum EEG patterns
Background: Manual annotation of seizures and interictal-ictal-injury continuum (IIIC) patterns in continuous EEG (cEEG) recorded from critically ill patients is a time-intensive process for clinicians and researchers. In this study, we evaluated the accuracy and efficiency of an automated clustering method to accelerate expert annotation of cEEG. New method: We learned a local dictionary from 97 ICU patients by applying k-medoids clustering to 592 features in the time and frequency domains. We utilized changepoint detection (CPD) to segment the cEEG recordings. We then computed a bag-of-words (BoW) representation for each segment. We further clustered the segments by affinity propagation. EEG experts scored the resulting clusters for each patient by labeling only the cluster medoids. We trained a random forest classifier to assess validity of the clusters. Results: Mean pairwise agreement of 62.6% using this automated method was not significantly different from interrater agreements using manual labeling (63.8%), demonstrating the validity of the method. We also found that it takes experts using our method 5.31 +/- 4.44 min to label the 30.19 +/- 3.84 h of cEEG data, more than 45 times faster than unaided manual review, demonstrating efficiency. Comparison with existing methods: Previous studies of EEG data labeling have generally yielded similar human expert interrater agreements, and lower agreements with automated methods. Conclusions: Our results suggest that long EEG recordings can be rapidly annotated by experts many times faster than unaided manual review through the use of an advanced clustering method
ManyDG: Many-domain Generalization for Healthcare Applications
The vast amount of health data has been continuously collected for each
patient, providing opportunities to support diverse healthcare predictive tasks
such as seizure detection and hospitalization prediction. Existing models are
mostly trained on other patients data and evaluated on new patients. Many of
them might suffer from poor generalizability. One key reason can be overfitting
due to the unique information related to patient identities and their data
collection environments, referred to as patient covariates in the paper. These
patient covariates usually do not contribute to predicting the targets but are
often difficult to remove. As a result, they can bias the model training
process and impede generalization. In healthcare applications, most existing
domain generalization methods assume a small number of domains. In this paper,
considering the diversity of patient covariates, we propose a new setting by
treating each patient as a separate domain (leading to many domains). We
develop a new domain generalization method ManyDG, that can scale to such
many-domain problems. Our method identifies the patient domain covariates by
mutual reconstruction and removes them via an orthogonal projection step.
Extensive experiments show that ManyDG can boost the generalization performance
on multiple real-world healthcare tasks (e.g., 3.7% Jaccard improvements on
MIMIC drug recommendation) and support realistic but challenging settings such
as insufficient data and continuous learning.Comment: The paper has been accepted by ICLR 2023, refer to
https://openreview.net/forum?id=lcSfirnflpW. We will release the data and
source codes here https://github.com/ycq091044/ManyD
Tools of Trade of the Next Blue-Collar Job? Antecedents, Design Features, and Outcomes of Interactive Labeling Systems
Supervised machine learning is becoming increasingly popular - and so is the need for annotated training data. Such data often needs to be manually labeled by human workers, not unlikely to negatively impact the involved workforce. To alleviate this issue, a new information systems class has emerged - interactive labeling systems. However, this young, but rapidly growing field lacks guidance and structure regarding the design of such systems. Against this backdrop, this paper describes antecedents, design features, and outcomes of interactive labeling systems. We perform a systematic literature review, identifying 188 relevant articles. Our results are presented as a morphological box with 14 dimensions, which we evaluate using card sorting. By additionally offering this box as a web-based artifact, we provide actionable guidance for interactive labeling system development for scholars and practitioners. Lastly, we discuss imbalances in the article distribution of our morphological box and suggest future work directions