877 research outputs found
Convolutional neural network for breathing phase detection in lung sounds
We applied deep learning to create an algorithm for breathing phase detection
in lung sound recordings, and we compared the breathing phases detected by the
algorithm and manually annotated by two experienced lung sound researchers. Our
algorithm uses a convolutional neural network with spectrograms as the
features, removing the need to specify features explicitly. We trained and
evaluated the algorithm using three subsets that are larger than previously
seen in the literature. We evaluated the performance of the method using two
methods. First, discrete count of agreed breathing phases (using 50% overlap
between a pair of boxes), shows a mean agreement with lung sound experts of 97%
for inspiration and 87% for expiration. Second, the fraction of time of
agreement (in seconds) gives higher pseudo-kappa values for inspiration
(0.73-0.88) than expiration (0.63-0.84), showing an average sensitivity of 97%
and an average specificity of 84%. With both evaluation methods, the agreement
between the annotators and the algorithm shows human level performance for the
algorithm. The developed algorithm is valid for detecting breathing phases in
lung sound recordings
DeepCough: A Deep Convolutional Neural Network in A Wearable Cough Detection System
In this paper, we present a system that employs a wearable acoustic sensor
and a deep convolutional neural network for detecting coughs. We evaluate the
performance of our system on 14 healthy volunteers and compare it to that of
other cough detection systems that have been reported in the literature.
Experimental results show that our system achieves a classification sensitivity
of 95.1% and a specificity of 99.5%.Comment: BioCAS-201
Feature Detection in Medical Images Using Deep Learning
This project explores the use of deep learning to predict age based on pediatric hand X-Rays. Data from the Radiological Society of North America’s pediatric bone age challenge were used to train and evaluate a convolutional neural network. The project used InceptionV3, a CNN developed by Google, that was pre-trained on ImageNet, a popular online image dataset. Our fine-tuned version of InceptionV3 yielded an average error of less than 10 months between predicted and actual age. This project shows the effectiveness of deep learning in analyzing medical images and the potential for even greater improvements in the future. In addition to the technological and potential clinical benefits of these methods, this project will serve as a useful pedagogical tool for introducing the challenges and applications of deep learning to the Bryant community
2D respiratory sound analysis to detect lung abnormalities
In this paper, we analyze deep visual features from 2D data representation(s) of the respiratory sound to detect evidence of lung abnormalities. The primary motivation behind this is that visual cues are more important in decision-making than raw data (lung sound). Early detection and prompt treatments are essential for any future possible respiratory disorders, and respiratory sound is proven to be one of the biomarkers. In contrast to state-of-the-art approaches, we aim at understanding/analyzing visual features using our Convolutional Neural Networks (CNN) tailored Deep Learning Models, where we consider all possible 2D data such as Spectrogram, Mel-frequency Cepstral Coefficients (MFCC), spectral centroid, and spectral roll-off. In our experiments, using the publicly available respiratory sound database named ICBHI 2017 (5.5 hours of recordings containing 6898 respiratory cycles from 126 subjects), we received the highest performance with the area under the curve of 0.79 from Spectrogram as opposed to 0.48 AUC from the raw data from a pre-trained deep learning model: VGG16. We also used machine learning algorithms using reliable data to improve Our study proved that 2D data representation could help better understand/analyze lung abnormalities as compared to 1D data. Our findings are also contrasted with those of earlier studies. For purposes of generality, we used the MFCC of neutrinos to determine if picture data or raw data produced superior results
NRC-Net: Automated noise robust cardio net for detecting valvular cardiac diseases using optimum transformation method with heart sound signals
Cardiovascular diseases (CVDs) can be effectively treated when detected
early, reducing mortality rates significantly. Traditionally, phonocardiogram
(PCG) signals have been utilized for detecting cardiovascular disease due to
their cost-effectiveness and simplicity. Nevertheless, various environmental
and physiological noises frequently affect the PCG signals, compromising their
essential distinctive characteristics. The prevalence of this issue in
overcrowded and resource-constrained hospitals can compromise the accuracy of
medical diagnoses. Therefore, this study aims to discover the optimal
transformation method for detecting CVDs using noisy heart sound signals and
propose a noise robust network to improve the CVDs classification
performance.For the identification of the optimal transformation method for
noisy heart sound data mel-frequency cepstral coefficients (MFCCs), short-time
Fourier transform (STFT), constant-Q nonstationary Gabor transform (CQT) and
continuous wavelet transform (CWT) has been used with VGG16. Furthermore, we
propose a novel convolutional recurrent neural network (CRNN) architecture
called noise robust cardio net (NRC-Net), which is a lightweight model to
classify mitral regurgitation, aortic stenosis, mitral stenosis, mitral valve
prolapse, and normal heart sounds using PCG signals contaminated with
respiratory and random noises. An attention block is included to extract
important temporal and spatial features from the noisy corrupted heart
sound.The results of this study indicate that,CWT is the optimal transformation
method for noisy heart sound signals. When evaluated on the GitHub heart sound
dataset, CWT demonstrates an accuracy of 95.69% for VGG16, which is 1.95%
better than the second-best CQT transformation technique. Moreover, our
proposed NRC-Net with CWT obtained an accuracy of 97.4%, which is 1.71% higher
than the VGG16
- …