17 research outputs found

    Can Machine Learning Be Used to Recognize and Diagnose Coughs?

    Full text link
    Emerging wireless technologies, such as 5G and beyond, are bringing new use cases to the forefront, one of the most prominent being machine learning empowered health care. One of the notable modern medical concerns that impose an immense worldwide health burden are respiratory infections. Since cough is an essential symptom of many respiratory infections, an automated system to screen for respiratory diseases based on raw cough data would have a multitude of beneficial research and medical applications. In literature, machine learning has already been successfully used to detect cough events in controlled environments. In this paper, we present a low complexity, automated recognition and diagnostic tool for screening respiratory infections that utilizes Convolutional Neural Networks (CNNs) to detect cough within environment audio and diagnose three potential illnesses (i.e., bronchitis, bronchiolitis and pertussis) based on their unique cough audio features. Both proposed detection and diagnosis models achieve an accuracy of over 89%, while also remaining computationally efficient. Results show that the proposed system is successfully able to detect and separate cough events from background noise. Moreover, the proposed single diagnosis model is capable of distinguishing between different illnesses without the need of separate models.Comment: Accepted in IEEE International Conference on E-Health and Bioengineering - EHB 202

    Deep recurrent neural networks with attention mechanisms for respiratory anomaly classification.

    Get PDF
    In recent years, a variety of deep learning techniques and methods have been adopted to provide AI solutions to issues within the medical field, with one specific area being audio-based classification of medical datasets. This research aims to create a novel deep learning architecture for this purpose, with a variety of different layer structures implemented for undertaking audio classification. Specifically, bidirectional Long Short-Term Memory (BiLSTM) and Gated Recurrent Units (GRU) networks in conjunction with an attention mechanism, are implemented in this research for chronic and non-chronic lung disease and COVID-19 diagnosis. We employ two audio datasets, i.e. the Respiratory Sound and the Coswara datasets, to evaluate the proposed model architectures pertaining to lung disease classification. The Respiratory Sound Database contains audio data with respect to lung conditions such as Chronic Obstructive Pulmonary Disease (COPD) and asthma, while the Coswara dataset contains coughing audio samples associated with COVID-19. After a comprehensive evaluation and experimentation process, as the most performant architecture, the proposed attention BiLSTM network (A-BiLSTM) achieves accuracy rates of 96.2% and 96.8% for the Respiratory Sound and the Coswara datasets, respectively. Our research indicates that the implementation of the BiLSTM and attention mechanism was effective in improving performance for undertaking audio classification with respect to various lung condition diagnoses

    Сверточная нейронная сеть для ИТ-диагностики легких

    Get PDF
    Предметом исследований является использовании технологии обработки голоса пациента в ИТ- медицине. Цель статьи – разработать нейронную сеть для диагностики заболеваний легких с помощью звукового анализа голоса пациента. Исследование включает в себя обучение нейронной сети, разработку мобильной программы для сбора звука пациента, извлечение звуковых характеристик на стороне сервера, диагностику звуковых данных с использованием обученной нейронной сети и возврат результатов диагностики в мобильную программу приложения. Представлена блок-схема обработки голоса от исходного сигнала до извлечения аудиофайла, в качестве примера приведено извлечение функций MFCC и FBank. Приведена структура сверточной нейронной сети (CNN), которая была обучена на стандарном наборе данных респираторных заболеваний. Приведен упрощенный процесс классификации звуков дыхания, необходимых для прогнозирования заболеваний легких. Для практической реализации использована в среде программирования Pyton сеть VGGish, которая имеет сетевые параметры, обученные с помощью набора данных. Эксприменты проведены на платформе Android service framework, которая разделена на две части: Android front-end и серверную. Интерфейсная часть реализует интерактивную функцию пользователя и отвечает за ввод аудиоданных. После загрузки аудио сервер выполнит предварительную обработку аудио, и вызовет CNN для классификации аудио, результаты возвращаются во внешний модуль на смартфоне. Лучшая точность модели достигла 83,6 %

    Classification of phonocardiograms with convolutional neural networks

    Get PDF
    The diagnosis of heart diseases from heart sounds is a matter of many years. This is the effect of having too many people with heart diseases in the world. Studies on heart sounds are usually based on classification for helping doctors. In other words, these studies are a substructure of clinical decision support systems. In this study, three different heart sound data in the PASCAL Btraining data set such as normal, murmur, and extrasystole are classified. Phonocardiograms which were obtained from heart sounds in the data set were used for classification. Both Artificial Neural Network (ANN) and Convolutional Neural Network (CNN) were used for classification to compare obtained results. In these studies, the obtained results show that the CNN classification gives the better result with 97.9% classification accuracy according to the results of ANN. Thus, CNN emerges as the ideal classification tool for the classification of heart sounds with variable characteristics

    Lung Sounds Classification Based on Time Domain Features

    Get PDF
    Signal complexity in lung sounds is assumed to be able to differentiate and classify characteristic lung sound between normal and abnormal in most cases. Previous research has employed a variety of modification approaches to obtain lung sound features. In contrast to earlier research, time-domain features were used to extract features in lung sound classification. Electromyogram (EMG) signal analysis frequently employs this time-domain characteristic. Time-domain features are MAV, SSI, Var, RMS, LOG, WL, AAC, DASDV, and AFB. The benefit of this method is that it allows for direct feature extraction without the requirement for transformation. Several classifiers were used to examine five different types of lung sound data. The highest accuracy was 93.9 percent, obtained Using the decision tree with 9 types of time-domain features. The proposed method could extract features from lung sounds as an alternative

    Benchmarking of eight recurrent neural network variants for breath phase and adventitious sound detection on a self-developed open-access lung sound database-HF_Lung_V1

    Full text link
    A reliable, remote, and continuous real-time respiratory sound monitor with automated respiratory sound analysis ability is urgently required in many clinical scenarios-such as in monitoring disease progression of coronavirus disease 2019-to replace conventional auscultation with a handheld stethoscope. However, a robust computerized respiratory sound analysis algorithm has not yet been validated in practical applications. In this study, we developed a lung sound database (HF_Lung_V1) comprising 9,765 audio files of lung sounds (duration of 15 s each), 34,095 inhalation labels, 18,349 exhalation labels, 13,883 continuous adventitious sound (CAS) labels (comprising 8,457 wheeze labels, 686 stridor labels, and 4,740 rhonchi labels), and 15,606 discontinuous adventitious sound labels (all crackles). We conducted benchmark tests for long short-term memory (LSTM), gated recurrent unit (GRU), bidirectional LSTM (BiLSTM), bidirectional GRU (BiGRU), convolutional neural network (CNN)-LSTM, CNN-GRU, CNN-BiLSTM, and CNN-BiGRU models for breath phase detection and adventitious sound detection. We also conducted a performance comparison between the LSTM-based and GRU-based models, between unidirectional and bidirectional models, and between models with and without a CNN. The results revealed that these models exhibited adequate performance in lung sound analysis. The GRU-based models outperformed, in terms of F1 scores and areas under the receiver operating characteristic curves, the LSTM-based models in most of the defined tasks. Furthermore, all bidirectional models outperformed their unidirectional counterparts. Finally, the addition of a CNN improved the accuracy of lung sound analysis, especially in the CAS detection tasks.Comment: 48 pages, 8 figures. To be submitte

    Automatic classification of adventitious respiratory sounds: a (un)solved problem?

    Get PDF
    (1) Background: Patients with respiratory conditions typically exhibit adventitious respiratory sounds (ARS), such as wheezes and crackles. ARS events have variable duration. In this work we studied the influence of event duration on automatic ARS classification, namely, how the creation of the Other class (negative class) affected the classifiers’ performance. (2) Methods: We conducted a set of experiments where we varied the durations of the other events on three tasks: crackle vs. wheeze vs. other (3 Class); crackle vs. other (2 Class Crackles); and wheeze vs. other (2 Class Wheezes). Four classifiers (linear discriminant analysis, support vector machines, boosted trees, and convolutional neural networks) were evaluated on those tasks using an open access respiratory sound database. (3) Results: While on the 3 Class task with fixed durations, the best classifier achieved an accuracy of 96.9%, the same classifier reached an accuracy of 81.8% on the more realistic 3 Class task with variable durations. (4) Conclusion: These results demonstrate the importance of experimental design on the assessment of the performance of automatic ARS classification algorithms. Furthermore, they also indicate, unlike what is stated in the literature, that the automatic classification of ARS is not a solved problem, as the algorithms’ performance decreases substantially under complex evaluation scenarios.publishe
    corecore