12,616 research outputs found
Deep learning with convolutional neural networks for decoding and visualization of EEG pathology
We apply convolutional neural networks (ConvNets) to the task of
distinguishing pathological from normal EEG recordings in the Temple University
Hospital EEG Abnormal Corpus. We use two basic, shallow and deep ConvNet
architectures recently shown to decode task-related information from EEG at
least as well as established algorithms designed for this purpose. In decoding
EEG pathology, both ConvNets reached substantially better accuracies (about 6%
better, ~85% vs. ~79%) than the only published result for this dataset, and
were still better when using only 1 minute of each recording for training and
only six seconds of each recording for testing. We used automated methods to
optimize architectural hyperparameters and found intriguingly different ConvNet
architectures, e.g., with max pooling as the only nonlinearity. Visualizations
of the ConvNet decoding behavior showed that they used spectral power changes
in the delta (0-4 Hz) and theta (4-8 Hz) frequency range, possibly alongside
other features, consistent with expectations derived from spectral analysis of
the EEG data and from the textual medical reports. Analysis of the textual
medical reports also highlighted the potential for accuracy increases by
integrating contextual information, such as the age of subjects. In summary,
the ConvNets and visualization techniques used in this study constitute a next
step towards clinically useful automated EEG diagnosis and establish a new
baseline for future work on this topic.Comment: Published at IEEE SPMB 2017 https://www.ieeespmb.org/2017
Articulatory and bottleneck features for speaker-independent ASR of dysarthric speech
The rapid population aging has stimulated the development of assistive
devices that provide personalized medical support to the needies suffering from
various etiologies. One prominent clinical application is a computer-assisted
speech training system which enables personalized speech therapy to patients
impaired by communicative disorders in the patient's home environment. Such a
system relies on the robust automatic speech recognition (ASR) technology to be
able to provide accurate articulation feedback. With the long-term aim of
developing off-the-shelf ASR systems that can be incorporated in clinical
context without prior speaker information, we compare the ASR performance of
speaker-independent bottleneck and articulatory features on dysarthric speech
used in conjunction with dedicated neural network-based acoustic models that
have been shown to be robust against spectrotemporal deviations. We report ASR
performance of these systems on two dysarthric speech datasets of different
characteristics to quantify the achieved performance gains. Despite the
remaining performance gap between the dysarthric and normal speech, significant
improvements have been reported on both datasets using speaker-independent ASR
architectures.Comment: to appear in Computer Speech & Language -
https://doi.org/10.1016/j.csl.2019.05.002 - arXiv admin note: substantial
text overlap with arXiv:1807.1094
Deep Neural Networks for the Recognition and Classification of Heart Murmurs Using Neuromorphic Auditory Sensors
Auscultation is one of the most used techniques for
detecting cardiovascular diseases, which is one of the main causes
of death in the world. Heart murmurs are the most common abnormal
finding when a patient visits the physician for auscultation.
These heart sounds can either be innocent, which are harmless, or
abnormal, which may be a sign of a more serious heart condition.
However, the accuracy rate of primary care physicians and expert
cardiologists when auscultating is not good enough to avoid most
of both type-I (healthy patients are sent for echocardiogram) and
type-II (pathological patients are sent home without medication or
treatment) errors made. In this paper, the authors present a novel
convolutional neural network based tool for classifying between
healthy people and pathological patients using a neuromorphic
auditory sensor for FPGA that is able to decompose the audio into
frequency bands in real time. For this purpose, different networks
have been trained with the heart murmur information contained in
heart sound recordings obtained from nine different heart sound
databases sourced from multiple research groups. These samples
are segmented and preprocessed using the neuromorphic auditory
sensor to decompose their audio information into frequency
bands and, after that, sonogram images with the same size are
generated. These images have been used to train and test different
convolutional neural network architectures. The best results
have been obtained with a modified version of the AlexNet model,
achieving 97% accuracy (specificity: 95.12%, sensitivity: 93.20%,
PhysioNet/CinC Challenge 2016 score: 0.9416). This tool could aid
cardiologists and primary care physicians in the auscultation process,
improving the decision making task and reducing type-I and
type-II errors.Ministerio de Economía y Competitividad TEC2016-77785-
- …