6 research outputs found

    Advanced Human Activity Recognition through Data Augmentation and Feature Concatenation of Micro-Doppler Signatures

    Get PDF
    Developing accurate classification models for radar-based Human Activity Recognition (HAR), capable of solving real-world problems, depends heavily on the amount of available data. In this paper, we propose a simple, effective, and generalizable data augmentation strategy along with preprocessing for micro-Doppler signatures to enhance recognition performance. By leveraging the decomposition properties of the Discrete Wavelet Transform (DWT), new samples are generated with distinct characteristics that do not overlap with those of the original samples. The micro-Doppler signatures are projected onto the DWT space for the decomposition process using the Haar wavelet. The returned decomposition components are used in different configurations to generate new data. Three new samples are obtained from a single spectrogram, which increases the amount of training data without creating duplicates. Next, the augmented samples are processed using the Sobel filter. This step allows each sample to be expanded into three representations, including the gradient in the x-direction (Dx), y-direction (Dy), and both x- and y-directions (Dxy). These representations are used as input for training a three-input convolutional neural network-long short-term memory support vector machine (CNN-LSTM-SVM) model. We have assessed the feasibility of our solution by evaluating it on three datasets containing micro-Doppler signatures of human activities, including Frequency Modulated Continuous Wave (FMCW) 77 GHz, FMCW 24 GHz, and Impulse Radio Ultra-Wide Band (IR-UWB) 10 GHz datasets. Several experiments have been carried out to evaluate the model\u27s performance with the inclusion of additional samples. The model was trained from scratch only on the augmented samples and tested on the original samples. Our augmentation approach has been thoroughly evaluated using various metrics, including accuracy, precision, recall, and F1-score. The results demonstrate a substantial improvement in the recognition rate and effectively alleviate the overfitting effect. Accuracies of 96.47%, 94.27%, and 98.18% are obtained for the FMCW 77 GHz, FMCW 24 GHz, and IR- UWB 10 GHz datasets, respectively. The findings of the study demonstrate the utility of DWT to enrich micro-Doppler training samples to improve HAR performance. Furthermore, the processing step was found to be efficient in enhancing the classification accuracy, achieving 96.78%, 96.32%, and 100% for the FMCW 77 GHz, FMCW 24 GHz, and IR-UWB 10 GHz datasets, respectively

    A Study on the Generation of Synthetic Data using Generative Adversarial Network for Optimization of Diagnostic Performance on Paranasal Sinusitis Radiography

    Get PDF
    λ”₯λŸ¬λ‹ κΈ°μˆ μ€ μ΅œκ·Όλ“€μ–΄ λ‹€μ–‘ν•œ 의료 뢄야에 적용되고 μžˆλ‹€. λ”₯λŸ¬λ‹ λͺ¨λΈμ˜ 효과적인 ν•™μŠ΅μ„ μœ„ν•΄μ„œλŠ” λ°©λŒ€ν•œ μ–‘μ˜ 데이터가 ν•„μš”ν•œλ°, μ‹€μ œλ‘œ ν™˜μžκ°€ μ˜λ£ŒκΈ°κ΄€μ—μ„œ 받은 검사λ₯Ό 톡해 μ–»μ–΄μ§€λŠ” 의료 λ°μ΄ν„°μ˜ νŠΉμ„±μƒ 데이터가 λΆ€μ‘±ν•œ κ²½μš°κ°€ 많으며, μ΄λŠ” λ”₯λŸ¬λ‹ κΈ°μˆ μ„ 의료 뢄야에 μ μš©ν•¨μ— μžˆμ–΄ 큰 ν•œκ³„μ μ΄λ‹€. μ΄λŸ¬ν•œ ν•™μŠ΅ 데이터 뢀쑱을 κ·Ήλ³΅ν•˜κΈ° μœ„ν•œ λ‹€μ–‘ν•œ 연ꡬ듀이 ν˜„μž¬κΉŒμ§€λ„ ν™œλ°œνžˆ μ§„ν–‰λ˜κ³  μžˆλ‹€. λ³Έ μ—°κ΅¬μ—μ„œλŠ” ν•™μŠ΅μš© 의료 데이터가 λΆ€μ‘±ν•œ μƒν™©μ—μ„œ μš°μˆ˜ν•œ νŒλ… μ„±λŠ₯에 도달할 수 μžˆλ„λ‘ 생성적 μ λŒ€ 신경망 (Generative Adversarial Network, GAN)을 μ΄μš©ν•˜μ—¬ ν•©μ„± 데이터 증강 (data augmentation)을 μˆ˜ν–‰ν•˜κ³  κ°€μž₯ μš°μˆ˜ν•œ μ„±λŠ₯에 μ΅œμ ν™”λœ λͺ¨λΈμ„ μžλ™μ μœΌλ‘œ μ°Ύμ•„ 쀄 수 μžˆλŠ” 방법을 μ œμ•ˆν•œλ‹€. 기쑴의 κ΄€λ ¨ μ—°κ΅¬μ—μ„œλŠ” GAN λͺ¨λΈμ„ μ΄μš©ν•˜μ—¬ ν•©μ„± 데이터λ₯Ό μƒμ„±ν•˜κ³  데이터 증강 κ°œμˆ˜κ°€ ν•™μŠ΅ λͺ¨λΈμ˜ μ„±λŠ₯에 λ―ΈμΉ˜λŠ” 영ν–₯에 λŒ€ν•œ μ‹€ν—˜μ„ μ§„ν–‰ν•˜μ˜€λ‹€. μ•žμ„  μ—°κ΅¬λŠ” 데이터 증강 μƒ˜ν”Œμ„ μΌμ •ν•œ 간격없이 μž„μ˜μ μœΌλ‘œ μ¦κ°€μ‹œμΌ°κΈ° λ•Œλ¬Έμ— μ •λŸ‰μ μ΄μ§€ μ•Šλ‹€λŠ” 점과 μ μ ˆν•œ 양을 νƒμƒ‰ν•˜κΈ° μœ„ν•΄ 증강 개수λ₯Ό 맀번 μ‘°μ •ν•΄μ•Όλ˜λŠ” μˆ˜λ™ μž‘μ—…μ΄ κ°œμž…λ˜λ―€λ‘œ μƒλ‹Ήν•œ μ‹œκ°„μ΄ μ†Œμš”λœλ‹€λŠ” ν•œκ³„μ μ΄ μžˆλ‹€. λ³Έ μ—°κ΅¬μ—μ„œλŠ” GAN 기반의 ν•©μ„± 데이터 증강을 μˆ˜ν–‰ν•˜λŠ” κ³Όμ •μ—μ„œ μ–΄λŠ μ‹œμ μ—μ„œ κ°€μž₯ μ΅œμ ν™”λœ λͺ¨λΈ μ„±λŠ₯이 λ‚˜νƒ€λ‚˜λŠ”μ§€μ— λŒ€ν•˜μ—¬ μžλ™ν™” κ°œλ…μœΌλ‘œ μ ‘κ·Όν•œλ‹€. 연ꡬ λ°μ΄ν„°λŠ” 2μ°¨ λ³‘μ›μ—μ„œ μˆ˜μ§‘ν•œ 뢀비동 (Paranasal Sinuses, PNS) X-ray 이미지와 μΌλ°˜μ„±μ„ ν™•λ³΄ν•˜κΈ° μœ„ν•΄ μ™ΈλΆ€ 검증 (external validation)용으둜 3μ°¨ λ³‘μ›μ—μ„œ μˆ˜μ§‘ν•œ PNS X-ray μ΄λ―Έμ§€λ‘œ κ΅¬μ„±λ˜μ–΄ μžˆλ‹€. μ‹€ν—˜μ— μ‚¬μš©λ  λͺ¨λ“  λ°μ΄ν„°λŠ” μ „μ²˜λ¦¬ κ³Όμ •μ—μ„œ 상악동 (maxillary sinus)λΆ€μœ„λ₯Ό κΈ°μ€€μœΌλ‘œ μž˜λΌλ‚΄μ–΄ (cropping) μ˜μƒ 패치λ₯Ό μƒμ„±ν•˜μ˜€λ‹€. 데이터 증강 기법은 두 κ°€μ§€μ˜ λ°©λ²•μœΌλ‘œ μˆ˜ν–‰ν•˜μ˜€λ‹€. 첫 번째 방법은 전톡적 데이터 증강 기법 (Conventional data augmentation)μ—μ„œ ν”νžˆ μ‚¬μš©λ˜λŠ” νšŒμ „ (-10λ„μ—μ„œ 10도 λ²”μœ„ λ‚΄λ‘œ)κ³Ό 쒌우 뒀집기 (horizontal flipping) λ₯Ό μ μš©ν•˜μ˜€κ³ , 두 번째 λ°©λ²•μœΌλ‘œλŠ” GAN을 μ΄μš©ν•˜μ—¬ ν•©μ„± 데이터λ₯Ό 생성 ν›„ 원본 데이터와 κ²°ν•©μ‹œμΌœ 증강을 ν•˜λŠ” ν•©μ„± 데이터 증강 기법이닀. GAN기반의 ν•©μ„± 데이터 증강을 μˆ˜ν–‰ν•˜κΈ° μœ„ν•΄μ„œλŠ” Auxiliary Classifier GAN (ACGAN) λͺ¨λΈμ„ κΈ°μ‘΄ ν•™μŠ΅μš© 데이터에 μ μš©ν•˜μ—¬ μƒˆλ‘œμš΄ 데이터λ₯Ό μƒμ„±ν•˜μ˜€λ‹€. λ³Έ 과정을 톡해 μƒμ„±λœ 데이터λ₯Ό μΌμ •ν•œ 배율둜 μ¦κ°€μ‹œν‚€λ©°, 이λ₯Ό λ°”νƒ•μœΌλ‘œ ν•™μŠ΅λœ λ”₯λŸ¬λ‹ λͺ¨λΈμ΄ κ°€μž₯ 높은 μ„±λŠ₯을 λ³΄μ΄λŠ” 졜적의 배율 (optimal multiples)을 λ°œκ²¬ν•  수 μžˆλ„λ‘ μžλ™ν™” 방법을 κ΅¬μΆ•ν•˜μ˜€λ‹€. 이미지 λΆ„λ₯˜ κ³Όμ •μ—μ„œλŠ” CheXNet λͺ¨λΈμ„ μ΄μš©ν•˜μ—¬ 전이 ν•™μŠ΅ (transfer learning)을 μˆ˜ν–‰ν•˜μ˜€κ³  이후, λ³Έ μ—°κ΅¬μ˜ λͺ¨λΈ μ„±λŠ₯을 ν‰κ°€ν•˜κΈ° μœ„ν•΄μ„œ 전톡적 데이터 증강 기법을 μ΄μš©ν•˜μ—¬ ν•™μŠ΅ν•œ λͺ¨λΈμ˜ μ„±λŠ₯κ³Ό GAN 기반의 ν•©μ„± 데이터 증강 κΈ°λ²•μ˜ 졜적의 증강 배율둜 ν•™μŠ΅ν•œ λͺ¨λΈμ˜ μ„±λŠ₯을 λΉ„κ΅ν•˜μ˜€λ‹€. λ”λΆˆμ–΄ 두 λͺ¨λΈμ˜ μ„±λŠ₯ κ°„μ˜ 톡계적 μœ μ˜μ„±μ„ μ œμ‹œν•˜μ˜€λ‹€. λͺ¨λΈ μ„±λŠ₯ 평가 κ²°κ³Ό, λ³Έ μ—°κ΅¬μ—μ„œ μ œμ•ˆν•œ GAN 기반의 데이터 증강 기법 λͺ¨λΈμ΄ 전톡적 데이터 증강 기법보닀 λ‚΄λΆ€ 및 μ™ΈλΆ€ κ²€μ¦μ—μ„œ λͺ¨λ‘ μš°μˆ˜ν•œ μ„±λŠ₯을 λ³΄μ˜€μŒμ„ μž…μ¦ν•  수 μžˆμ—ˆλ‹€. λ³Έ μ—°κ΅¬μ—μ„œ μ œμ‹œν•œ GAN 기반의 데이터 증강 기법은 X-ray μ˜μƒ 뿐만 μ•„λ‹ˆλΌ λ‹€μ–‘ν•œ 의료 μ˜μƒμ—μ„œλ„ 적용될 수 있으며, μ¦κ°•μ‹œν‚¬ 데이터 μ–‘μ˜ 크기λ₯Ό μˆ˜λ™μœΌλ‘œ μ„€μ •ν•΄μ£ΌλŠ” 이전 λ°©μ‹κ³ΌλŠ” 달리 μΌμ •ν•œ 배율둜 μ¦κ°€μ‹œν‚€λ©° κ°€μž₯ μš°μˆ˜ν•œ μ„±λŠ₯에 μ΅œμ ν™”λœ λͺ¨λΈ 및 증강 λ°°μœ¨μ„ μžλ™μ μœΌλ‘œ μ°Ύμ•„μ£ΌκΈ° λ•Œλ¬Έμ— νŽΈλ¦¬μ„±κ³Ό 정확성을 μ œκ³΅ν•  것이닀. μΆ”ν›„ 데이터 증강 기법 κ΄€λ ¨ μ—°κ΅¬μ—μ„œ GAN 기반의 ν•©μ„± 데이터 증강 μžλ™ν™” μ•Œκ³ λ¦¬μ¦˜μ΄ 더 ν™œμš©λ  κ²ƒμœΌλ‘œ κΈ°λŒ€ν•œλ‹€.Recently, rapid development of deep learning has led to a state-of-the-art performance in a wide range of computer vision tasks, mainly through large-scale datasets. However, acquiring a large amount of dataset is restricted and challenging in medical domain. This study aimed to investigate the feasibility of synthetic data augmentation using generative adversarial networks (GAN) by suggesting an automatic method to determine the ideal amount of data augmentation in order to develop deep learning-based model optimized for best performance in a limited datasets. This retrospective study evaluated the Water’s view radiographs of 250 patients diagnosed with chronic sinusitis, who underwent radiographic examination, between January 2010 and January 2020. Image patches of 177 healthy sinuses and 212 maxillary sinusitis were created and randomly split into a training set (80%) and a test set (20%) to develop a deep learning model. During synthetic data augmentation process, Auxiliary Classifier GAN (ACGAN) was applied to generate synthetic images from the original training set. A method to determine the optimal amount of GAN-generated synthetic data to train the model with the highest performance was proposed herein. Transfer learning techniques were applied using a pre-trained CheXNet model for classification. The model was trained using conventional data augmentation and the proposed synthetic data augmentation. The performance of the models was evaluated based on the statistical analyses of the accuracy, sensitivity, specificity, positive and negative predictive values, area under the curve (AUC), and receiver operating characteristic, in both internal and external datasets. The experimental results verified that the proposed approach achieved a higher performance than conventional data augmentation, with an accuracy of 0.949, sensitivity of 0.917, specificity of 0.955, and an AUC of 0.964 using internal test set, and 0.917, 0.924, 0.909, and 0.955, respectively, using external test set. These values showed significant differences compared with the model trained using conventional data augmentation. The findings of this study suggest that the proposed deep learning approach could be used to assist radiologists and improve diagnosis with the use of deep learning technique even in a presence of lack of training data.제 1 μž₯ μ„œ λ‘  1 제 1 절 μ—°κ΅¬μ˜ λ°°κ²½ 1 제 2 절 κ΄€λ ¨ μ—°κ΅¬μ˜ 동ν–₯ 3 제 3 절 κ΄€λ ¨ μ—°κ΅¬μ˜ ν•œκ³„μ  5 제 4 절 μ—°κ΅¬μ˜ λͺ©μ  6 제 2 μž₯ 연ꡬ 방법 6 제 1 절 데이터 μˆ˜μ§‘ 및 ꡬ성 6 제 2 절 데이터 μ „μ²˜λ¦¬ κ³Όμ • 10 제 3 절 ACGAN λͺ¨λΈμ„ μ΄μš©ν•œ 데이터 생성 13 제 4 절 데이터 증강 기법 18 제 5 절 λΆ„λ₯˜ λͺ¨λΈ ν•™μŠ΅ κ³Όμ • 23 제 6 절 λͺ¨λΈ μ„±λŠ₯ 평가 및 μœ μ˜μ„± 검증 방법 24 제 3 μž₯ 연ꡬ κ²°κ³Ό 26 제 1 절 GAN 기반의 데이터 증강 κΈ°λ²•μ˜ 졜적 배율 27 제 2 절 λͺ¨λΈ μ„±λŠ₯ 평가 35 제 4 μž₯ κ³  μ°° 40 제 5 μž₯ κ²° λ‘  45 μ°Έκ³ λ¬Έν—Œ 46 Abstract 49석

    Trustworthy and Intelligent COVID-19 Diagnostic IoMT through XR and Deep-Learning-Based Clinic Data Access

    Get PDF
    This article presents a novel extended reality (XR) and deep-learning-based Internet-of-Medical-Things (IoMT) solution for the COVID-19 telemedicine diagnostic, which systematically combines virtual reality/augmented reality (AR) remote surgical plan/rehearse hardware, customized 5G cloud computing and deep learning algorithms to provide real-time COVID-19 treatment scheme clues. Compared to existing perception therapy techniques, our new technique can significantly improve performance and security. The system collected 25 clinic data from the 347 positive and 2270 negative COVID-19 patients in the Red Zone by 5G transmission. After that, a novel auxiliary classifier generative adversarial network-based intelligent prediction algorithm is conducted to train the new COVID-19 prediction model. Furthermore, The Copycat network is employed for the model stealing and attack for the IoMT to improve the security performance. To simplify the user interface and achieve an excellent user experience, we combined the Red Zone's guiding images with the Green Zone's view through the AR navigate clue by using 5G. The XR surgical plan/rehearse framework is designed, including all COVID-19 surgical requisite details that were developed with a real-time response guaranteed. The accuracy, recall, F1-score, and area under the ROC curve (AUC) area of our new IoMT were 0.92, 0.98, 0.95, and 0.98, respectively, which outperforms the existing perception techniques with significantly higher accuracy performance. The model stealing also has excellent performance, with the AUC area of 0.90 in Copycat slightly lower than the original model. This study suggests a new framework in the COVID-19 diagnostic integration and opens the new research about the integration of XR and deep learning for IoMT implementation

    Motion Classification Using Kinematically Sifted ACGAN-Synthesized Radar Micro-Doppler Signatures

    No full text
    corecore