6 research outputs found
Advanced Human Activity Recognition through Data Augmentation and Feature Concatenation of Micro-Doppler Signatures
Developing accurate classification models for radar-based Human Activity Recognition (HAR), capable of solving real-world problems, depends heavily on the amount of available data. In this paper, we propose a simple, effective, and generalizable data augmentation strategy along with preprocessing for micro-Doppler signatures to enhance recognition performance. By leveraging the decomposition properties of the Discrete Wavelet Transform (DWT), new samples are generated with distinct characteristics that do not overlap with those of the original samples. The micro-Doppler signatures are projected onto the DWT space for the decomposition process using the Haar wavelet. The returned decomposition components are used in different configurations to generate new data. Three new samples are obtained from a single spectrogram, which increases the amount of training data without creating duplicates. Next, the augmented samples are processed using the Sobel filter. This step allows each sample to be expanded into three representations, including the gradient in the x-direction (Dx), y-direction (Dy), and both x- and y-directions (Dxy). These representations are used as input for training a three-input convolutional neural network-long short-term memory support vector machine (CNN-LSTM-SVM) model. We have assessed the feasibility of our solution by evaluating it on three datasets containing micro-Doppler signatures of human activities, including Frequency Modulated Continuous Wave (FMCW) 77 GHz, FMCW 24 GHz, and Impulse Radio Ultra-Wide Band (IR-UWB) 10 GHz datasets. Several experiments have been carried out to evaluate the model\u27s performance with the inclusion of additional samples. The model was trained from scratch only on the augmented samples and tested on the original samples. Our augmentation approach has been thoroughly evaluated using various metrics, including accuracy, precision, recall, and F1-score. The results demonstrate a substantial improvement in the recognition rate and effectively alleviate the overfitting effect. Accuracies of 96.47%, 94.27%, and 98.18% are obtained for the FMCW 77 GHz, FMCW 24 GHz, and IR- UWB 10 GHz datasets, respectively. The findings of the study demonstrate the utility of DWT to enrich micro-Doppler training samples to improve HAR performance. Furthermore, the processing step was found to be efficient in enhancing the classification accuracy, achieving 96.78%, 96.32%, and 100% for the FMCW 77 GHz, FMCW 24 GHz, and IR-UWB 10 GHz datasets, respectively
A Study on the Generation of Synthetic Data using Generative Adversarial Network for Optimization of Diagnostic Performance on Paranasal Sinusitis Radiography
λ₯λ¬λ κΈ°μ μ μ΅κ·Όλ€μ΄ λ€μν μλ£ λΆμΌμ μ μ©λκ³ μλ€. λ₯λ¬λ λͺ¨λΈμ ν¨κ³Όμ μΈ νμ΅μ μν΄μλ λ°©λν μμ λ°μ΄ν°κ° νμνλ°, μ€μ λ‘ νμκ° μλ£κΈ°κ΄μμ λ°μ κ²μ¬λ₯Ό ν΅ν΄ μ»μ΄μ§λ μλ£ λ°μ΄ν°μ νΉμ±μ λ°μ΄ν°κ° λΆμ‘±ν κ²½μ°κ° λ§μΌλ©°, μ΄λ λ₯λ¬λ κΈ°μ μ μλ£ λΆμΌμ μ μ©ν¨μ μμ΄ ν° νκ³μ μ΄λ€. μ΄λ¬ν νμ΅ λ°μ΄ν° λΆμ‘±μ 극볡νκΈ° μν λ€μν μ°κ΅¬λ€μ΄ νμ¬κΉμ§λ νλ°ν μ§νλκ³ μλ€. λ³Έ μ°κ΅¬μμλ νμ΅μ© μλ£ λ°μ΄ν°κ° λΆμ‘±ν μν©μμ μ°μν νλ
μ±λ₯μ λλ¬ν μ μλλ‘ μμ±μ μ λ μ κ²½λ§ (Generative Adversarial Network, GAN)μ μ΄μ©νμ¬ ν©μ± λ°μ΄ν° μ¦κ° (data augmentation)μ μννκ³ κ°μ₯ μ°μν μ±λ₯μ μ΅μ νλ λͺ¨λΈμ μλμ μΌλ‘ μ°Ύμ μ€ μ μλ λ°©λ²μ μ μνλ€.
κΈ°μ‘΄μ κ΄λ ¨ μ°κ΅¬μμλ GAN λͺ¨λΈμ μ΄μ©νμ¬ ν©μ± λ°μ΄ν°λ₯Ό μμ±νκ³ λ°μ΄ν° μ¦κ° κ°μκ° νμ΅ λͺ¨λΈμ μ±λ₯μ λ―ΈμΉλ μν₯μ λν μ€νμ μ§ννμλ€. μμ μ°κ΅¬λ λ°μ΄ν° μ¦κ° μνμ μΌμ ν κ°κ²©μμ΄ μμμ μΌλ‘ μ¦κ°μμΌ°κΈ° λλ¬Έμ μ λμ μ΄μ§ μλ€λ μ κ³Ό μ μ ν μμ νμνκΈ° μν΄ μ¦κ° κ°μλ₯Ό λ§€λ² μ‘°μ ν΄μΌλλ μλ μμ
μ΄ κ°μ
λλ―λ‘ μλΉν μκ°μ΄ μμλλ€λ νκ³μ μ΄ μλ€.
λ³Έ μ°κ΅¬μμλ GAN κΈ°λ°μ ν©μ± λ°μ΄ν° μ¦κ°μ μννλ κ³Όμ μμ μ΄λ μμ μμ κ°μ₯ μ΅μ νλ λͺ¨λΈ μ±λ₯μ΄ λνλλμ§μ λνμ¬ μλν κ°λ
μΌλ‘ μ κ·Όνλ€. μ°κ΅¬ λ°μ΄ν°λ 2μ°¨ λ³μμμ μμ§ν λΆλΉλ (Paranasal Sinuses, PNS) X-ray μ΄λ―Έμ§μ μΌλ°μ±μ ν보νκΈ° μν΄ μΈλΆ κ²μ¦ (external validation)μ©μΌλ‘ 3μ°¨ λ³μμμ μμ§ν PNS X-ray μ΄λ―Έμ§λ‘ ꡬμ±λμ΄ μλ€. μ€νμ μ¬μ©λ λͺ¨λ λ°μ΄ν°λ μ μ²λ¦¬ κ³Όμ μμ μμ
λ (maxillary sinus)λΆμλ₯Ό κΈ°μ€μΌλ‘ μλΌλ΄μ΄ (cropping) μμ ν¨μΉλ₯Ό μμ±νμλ€.
λ°μ΄ν° μ¦κ° κΈ°λ²μ λ κ°μ§μ λ°©λ²μΌλ‘ μννμλ€. 첫 λ²μ§Έ λ°©λ²μ μ ν΅μ λ°μ΄ν° μ¦κ° κΈ°λ² (Conventional data augmentation)μμ νν μ¬μ©λλ νμ (-10λμμ 10λ λ²μ λ΄λ‘)κ³Ό μ’μ° λ€μ§κΈ° (horizontal flipping) λ₯Ό μ μ©νμκ³ , λ λ²μ§Έ λ°©λ²μΌλ‘λ GANμ μ΄μ©νμ¬ ν©μ± λ°μ΄ν°λ₯Ό μμ± ν μλ³Έ λ°μ΄ν°μ κ²°ν©μμΌ μ¦κ°μ νλ ν©μ± λ°μ΄ν° μ¦κ° κΈ°λ²μ΄λ€. GANκΈ°λ°μ ν©μ± λ°μ΄ν° μ¦κ°μ μννκΈ° μν΄μλ Auxiliary Classifier GAN (ACGAN) λͺ¨λΈμ κΈ°μ‘΄ νμ΅μ© λ°μ΄ν°μ μ μ©νμ¬ μλ‘μ΄ λ°μ΄ν°λ₯Ό μμ±νμλ€. λ³Έ κ³Όμ μ ν΅ν΄ μμ±λ λ°μ΄ν°λ₯Ό μΌμ ν λ°°μ¨λ‘ μ¦κ°μν€λ©°, μ΄λ₯Ό λ°νμΌλ‘ νμ΅λ λ₯λ¬λ λͺ¨λΈμ΄ κ°μ₯ λμ μ±λ₯μ 보μ΄λ μ΅μ μ λ°°μ¨ (optimal multiples)μ λ°κ²¬ν μ μλλ‘ μλν λ°©λ²μ ꡬμΆνμλ€.
μ΄λ―Έμ§ λΆλ₯ κ³Όμ μμλ CheXNet λͺ¨λΈμ μ΄μ©νμ¬ μ μ΄ νμ΅ (transfer learning)μ μννμκ³ μ΄ν, λ³Έ μ°κ΅¬μ λͺ¨λΈ μ±λ₯μ νκ°νκΈ° μν΄μ μ ν΅μ λ°μ΄ν° μ¦κ° κΈ°λ²μ μ΄μ©νμ¬ νμ΅ν λͺ¨λΈμ μ±λ₯κ³Ό GAN κΈ°λ°μ ν©μ± λ°μ΄ν° μ¦κ° κΈ°λ²μ μ΅μ μ μ¦κ° λ°°μ¨λ‘ νμ΅ν λͺ¨λΈμ μ±λ₯μ λΉκ΅νμλ€. λλΆμ΄ λ λͺ¨λΈμ μ±λ₯ κ°μ ν΅κ³μ μ μμ±μ μ μνμλ€.
λͺ¨λΈ μ±λ₯ νκ° κ²°κ³Ό, λ³Έ μ°κ΅¬μμ μ μν GAN κΈ°λ°μ λ°μ΄ν° μ¦κ° κΈ°λ² λͺ¨λΈμ΄ μ ν΅μ λ°μ΄ν° μ¦κ° κΈ°λ²λ³΄λ€ λ΄λΆ λ° μΈλΆ κ²μ¦μμ λͺ¨λ μ°μν μ±λ₯μ 보μμμ μ
μ¦ν μ μμλ€.
λ³Έ μ°κ΅¬μμ μ μν GAN κΈ°λ°μ λ°μ΄ν° μ¦κ° κΈ°λ²μ X-ray μμ λΏλ§ μλλΌ λ€μν μλ£ μμμμλ μ μ©λ μ μμΌλ©°, μ¦κ°μν¬ λ°μ΄ν° μμ ν¬κΈ°λ₯Ό μλμΌλ‘ μ€μ ν΄μ£Όλ μ΄μ λ°©μκ³Όλ λ¬λ¦¬ μΌμ ν λ°°μ¨λ‘ μ¦κ°μν€λ©° κ°μ₯ μ°μν μ±λ₯μ μ΅μ νλ λͺ¨λΈ λ° μ¦κ° λ°°μ¨μ μλμ μΌλ‘ μ°Ύμμ£ΌκΈ° λλ¬Έμ νΈλ¦¬μ±κ³Ό μ νμ±μ μ 곡ν κ²μ΄λ€. μΆν λ°μ΄ν° μ¦κ° κΈ°λ² κ΄λ ¨ μ°κ΅¬μμ GAN κΈ°λ°μ ν©μ± λ°μ΄ν° μ¦κ° μλν μκ³ λ¦¬μ¦μ΄ λ νμ©λ κ²μΌλ‘ κΈ°λνλ€.Recently, rapid development of deep learning has led to a state-of-the-art performance in a wide range of computer vision tasks, mainly through large-scale datasets. However, acquiring a large amount of dataset is restricted and challenging in medical domain.
This study aimed to investigate the feasibility of synthetic data augmentation using generative adversarial networks (GAN) by suggesting an automatic method to determine the ideal amount of data augmentation in order to develop deep learning-based model optimized for best performance in a limited datasets.
This retrospective study evaluated the Waterβs view radiographs of 250 patients diagnosed with chronic sinusitis, who underwent radiographic examination, between January 2010 and January 2020. Image patches of 177 healthy sinuses and 212 maxillary sinusitis were created and randomly split into a training set (80%) and a test set (20%) to develop a deep learning model. During synthetic data augmentation process, Auxiliary Classifier GAN (ACGAN) was applied to generate synthetic images from the original training set. A method to determine the optimal amount of GAN-generated synthetic data to train the model with the highest performance was proposed herein. Transfer learning techniques were applied using a pre-trained CheXNet model for classification. The model was trained using conventional data augmentation and the proposed synthetic data augmentation.
The performance of the models was evaluated based on the statistical analyses of the accuracy, sensitivity, specificity, positive and negative predictive values, area under the curve (AUC), and receiver operating characteristic, in both internal and external datasets. The experimental results verified that the proposed approach achieved a higher performance than conventional data augmentation, with an accuracy of 0.949, sensitivity of 0.917, specificity of 0.955, and an AUC of 0.964 using internal test set, and 0.917, 0.924, 0.909, and 0.955, respectively, using external test set. These values showed significant differences compared with the model trained using conventional data augmentation.
The findings of this study suggest that the proposed deep learning approach could be used to assist radiologists and improve diagnosis with the use of deep learning technique even in a presence of lack of training data.μ 1 μ₯ μ λ‘ 1
μ 1 μ μ°κ΅¬μ λ°°κ²½ 1
μ 2 μ κ΄λ ¨ μ°κ΅¬μ λν₯ 3
μ 3 μ κ΄λ ¨ μ°κ΅¬μ νκ³μ 5
μ 4 μ μ°κ΅¬μ λͺ©μ 6
μ 2 μ₯ μ°κ΅¬ λ°©λ² 6
μ 1 μ λ°μ΄ν° μμ§ λ° κ΅¬μ± 6
μ 2 μ λ°μ΄ν° μ μ²λ¦¬ κ³Όμ 10
μ 3 μ ACGAN λͺ¨λΈμ μ΄μ©ν λ°μ΄ν° μμ± 13
μ 4 μ λ°μ΄ν° μ¦κ° κΈ°λ² 18
μ 5 μ λΆλ₯ λͺ¨λΈ νμ΅ κ³Όμ 23
μ 6 μ λͺ¨λΈ μ±λ₯ νκ° λ° μ μμ± κ²μ¦ λ°©λ² 24
μ 3 μ₯ μ°κ΅¬ κ²°κ³Ό 26
μ 1 μ GAN κΈ°λ°μ λ°μ΄ν° μ¦κ° κΈ°λ²μ μ΅μ λ°°μ¨ 27
μ 2 μ λͺ¨λΈ μ±λ₯ νκ° 35
μ 4 μ₯ κ³ μ°° 40
μ 5 μ₯ κ²° λ‘ 45
μ°Έκ³ λ¬Έν 46
Abstract 49μ
Trustworthy and Intelligent COVID-19 Diagnostic IoMT through XR and Deep-Learning-Based Clinic Data Access
This article presents a novel extended reality (XR) and deep-learning-based Internet-of-Medical-Things (IoMT) solution for the COVID-19 telemedicine diagnostic, which systematically combines virtual reality/augmented reality (AR) remote surgical plan/rehearse hardware, customized 5G cloud computing and deep learning algorithms to provide real-time COVID-19 treatment scheme clues. Compared to existing perception therapy techniques, our new technique can significantly improve performance and security. The system collected 25 clinic data from the 347 positive and 2270 negative COVID-19 patients in the Red Zone by 5G transmission. After that, a novel auxiliary classifier generative adversarial network-based intelligent prediction algorithm is conducted to train the new COVID-19 prediction model. Furthermore, The Copycat network is employed for the model stealing and attack for the IoMT to improve the security performance. To simplify the user interface and achieve an excellent user experience, we combined the Red Zone's guiding images with the Green Zone's view through the AR navigate clue by using 5G. The XR surgical plan/rehearse framework is designed, including all COVID-19 surgical requisite details that were developed with a real-time response guaranteed. The accuracy, recall, F1-score, and area under the ROC curve (AUC) area of our new IoMT were 0.92, 0.98, 0.95, and 0.98, respectively, which outperforms the existing perception techniques with significantly higher accuracy performance. The model stealing also has excellent performance, with the AUC area of 0.90 in Copycat slightly lower than the original model. This study suggests a new framework in the COVID-19 diagnostic integration and opens the new research about the integration of XR and deep learning for IoMT implementation