86 research outputs found

    Deep Generative Modeling Based Retinal Image Analysis

    Get PDF
    In the recent past, deep learning algorithms have been widely used in retinal image analysis (fundus and OCT) to perform tasks like segmentation and classification. But to build robust and highly efficient deep learning models amount of the training images, the quality of the training images is extremely necessary. The quality of an image is also an extremely important factor for the clinical diagnosis of different diseases. The main aim of this thesis is to explore two relatively under-explored area of retinal image analysis, namely, the retinal image quality enhancement and artificial image synthesis. In this thesis, we proposed a series of deep generative modeling based algorithms to perform these above-mentioned tasks. From a mathematical perspective, the generative model is a statistical model of the joint probability distribution between an observable variable and a target variable. The generative adversarial network (GAN), variational auto-encoder(VAE) are some popular generative models. Generative models can be used to generate new samples from a given distribution. The OCT images have inherent speckle noise in it, fundus images do not suffer from noises in general, but the newly developed tele-ophthalmoscope devices produce images with relatively low spatial resolution and blur. Different GAN based algorithms were developed to generate corresponding high-quality images fro its low-quality counterpart. A combination of residual VAE and GAN was implemented to generate artificial retinal fundus images with their corresponding artificial blood vessel segmentation maps. This will not only help to generate new training images as many as needed but also will help to reduce the privacy issue of releasing personal medical data

    λ”₯λŸ¬λ‹μ„ μ΄μš©ν•œ λ…Ήλ‚΄μž₯ 진단 보쑰 μ‹œμŠ€ν…œ

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (박사) -- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ ν˜‘λ™κ³Όμ • λ°”μ΄μ˜€μ—”μ§€λ‹ˆμ–΄λ§μ „κ³΅, 2021. 2. 김희찬.λ³Έ λ…Όλ¬Έμ—μ„œλŠ” λ”₯ λŸ¬λ‹ 기반의 진단 보쑰 μ‹œμŠ€ν…œμ„ μ œμ•ˆν•˜μ˜€λ‹€. μƒˆλ‘œμš΄ 방법이 λ…Ήλ‚΄μž₯ 데이터에 μ μš©λ˜μ—ˆκ³  κ²°κ³Όλ₯Ό ν‰κ°€ν•˜μ˜€λ‹€. 첫번째 μ—°κ΅¬μ—μ„œλŠ” μŠ€νŽ™νŠΈλŸΌμ˜μ—­ λΉ›κ°„μ„­λ‹¨μΈ΅μ΄¬μ˜κΈ°(SD-OCT)λ₯Ό λ”₯ λŸ¬λ‹ λΆ„λ₯˜ κΈ°λ₯Ό μ΄μš©ν•΄ λΆ„μ„ν•˜μ˜€λ‹€. μŠ€νŽ™νŠΈλŸΌμ˜μ—­ λΉ›κ°„μ„­λ‹¨μΈ΅μ΄¬μ˜κΈ°λŠ” λ…Ήλ‚΄μž₯으둜 μΈν•œ ꡬ쑰적 손상을 ν‰κ°€ν•˜κΈ° μœ„ν•΄ μ‚¬μš©ν•˜λŠ” μž₯비이닀. λΆ„λ₯˜ μ•Œκ³ λ¦¬μ¦˜μ€ ν•©μ„± κ³± 신경망을 μ΄μš©ν•΄ 개발 λ˜μ—ˆμœΌλ©°, μŠ€νŽ™νŠΈλŸΌμ˜μ—­ λΉ›κ°„μ„­λ‹¨μΈ΅μ΄¬μ˜κΈ°μ˜ λ§λ§‰μ‹ κ²½μ„¬μœ μΈ΅(RNFL)κ³Ό ν™©λ°˜λΆ€ μ‹ κ²½μ ˆμ„Έν¬λ‚΄λ§μƒμΈ΅ (GCIPL) 사진을 μ΄μš©ν•΄ ν•™μŠ΅ν–ˆλ‹€. μ œμ•ˆν•œ 방법은 λ‘κ°œμ˜ 이미지λ₯Ό μž…λ ₯으둜 λ°›λŠ” μ΄μ€‘μž…λ ₯합성곱신경망(DICNN)이며, λ”₯ λŸ¬λ‹ λΆ„λ₯˜μ—μ„œ 효과적인 κ²ƒμœΌλ‘œ μ•Œλ €μ Έ μžˆλ‹€. μ΄μ€‘μž…λ ₯합성곱신경망은 λ§λ§‰μ‹ κ²½μ„¬μœ μΈ΅ κ³Ό μ‹ κ²½μ ˆμ„Έν¬μΈ΅ 의 λ‘κ»˜ 지도λ₯Ό μ΄μš©ν•˜μ—¬ ν•™μŠ΅ 됐으며, ν•™μŠ΅λœ λ„€νŠΈμ›Œν¬λŠ” λ…Ήλ‚΄μž₯κ³Ό 정상 ꡰ을 κ΅¬λΆ„ν•œλ‹€. μ΄μ€‘μž…λ ₯합성곱신경망은 정확도와 μˆ˜μ‹ κΈ°λ™μž‘νŠΉμ„±κ³‘μ„ ν•˜λ©΄μ  (AUC)으둜 평가 λ˜μ—ˆλ‹€. λ§λ§‰μ‹ κ²½μ„¬μœ μΈ΅κ³Ό μ‹ κ²½μ ˆμ„Έν¬μΈ΅ λ‘κ»˜ μ§€λ„λ‘œ ν•™μŠ΅λœ μ„€κ³„ν•œ λ”₯ λŸ¬λ‹ λͺ¨λΈμ„ μ‘°κΈ° λ…Ήλ‚΄μž₯κ³Ό 정상 ꡰ을 λΆ„λ₯˜ν•˜λŠ” μ„±λŠ₯을 ν‰κ°€ν•˜κ³  λΉ„κ΅ν•˜μ˜€λ‹€. μ„±λŠ₯평가 κ²°κ³Ό μ΄μ€‘μž…λ ₯합성곱신경망은 μ‘°κΈ° λ…Ήλ‚΄μž₯을 λΆ„λ₯˜ν•˜λŠ”데 0.869의 μˆ˜μ‹ κΈ°λ™μž‘νŠΉμ„±κ³‘μ„ μ˜λ„“μ΄μ™€ 0.921의 민감도, 0.756의 νŠΉμ΄λ„λ₯Ό λ³΄μ˜€λ‹€. λ‘λ²ˆμ§Έ μ—°κ΅¬μ—μ„œλŠ” λ”₯ λŸ¬λ‹μ„ μ΄μš©ν•΄ μ‹œμ‹ κ²½μœ λ‘μ‚¬μ§„μ˜ 해상도와 λŒ€λΉ„, 색감, 밝기λ₯Ό λ³΄μ •ν•˜λŠ” 방법을 μ œμ•ˆν•˜μ˜€λ‹€. μ‹œμ‹ κ²½μœ λ‘μ‚¬μ§„μ€ λ…Ήλ‚΄μž₯을 μ§„λ‹¨ν•˜λŠ”λ° μžˆμ–΄ 효과적인 κ²ƒμœΌλ‘œ μ•Œλ €μ Έ μžˆλ‹€. ν•˜μ§€λ§Œ, λ…Ήλ‚΄μž₯의 μ§„λ‹¨μ—μ„œ ν™˜μžμ˜ λ‚˜, μž‘μ€ 동곡, 맀체 뢈투λͺ…μ„± λ“±μœΌλ‘œ 인해 평가가 μ–΄λ €μš΄ κ²½μš°κ°€ μžˆλ‹€. 초 해상도와 보정 μ•Œκ³ λ¦¬μ¦˜μ€ 초 해상도 μ λŒ€μ μƒμ„±μ‹ κ²½λ§μ„ 톡해 κ°œλ°œλ˜μ—ˆλ‹€. 원본 κ³ ν•΄μƒλ„μ˜ μ‹œμ‹ κ²½ μœ λ‘ 사진은 저해상도 μ‚¬μ§„μœΌλ‘œ μΆ•μ†Œλ˜κ³ , λ³΄μ •λœ 고해상도 μ‹œμ‹ κ²½μœ λ‘μ‚¬μ§„μœΌλ‘œ 보정 되며, λ³΄μ •λœ 사진은 μ‹œμ‹ κ²½μ—¬λ°±μ˜ κ°€μ‹œμ„±κ³Ό 근처 ν˜ˆκ΄€μ„ 잘 보이도둝 ν›„μ²˜λ¦¬ μ•Œκ³ λ¦¬μ¦˜μ„ μ΄μš©ν•œλ‹€. 저해상도이미지λ₯Ό λ³΄μ •λœ κ³ ν•΄μƒλ„μ΄λ―Έμ§€λ‘œ λ³΅μ›ν•˜λŠ” 과정을 μ΄ˆν•΄μƒλ„μ λŒ€μ μ‹ κ²½λ§μ„ 톡해 ν•™μŠ΅ν•œλ‹€. μ„€κ³„ν•œ λ„€νŠΈμ›Œν¬λŠ” μ‹ ν˜Έ λŒ€ 작음 λΉ„(PSNR)κ³Ό κ΅¬μ‘°μ μœ μ‚¬μ„±(SSIM), 평균평가점(MOS)λ₯Ό μ΄μš©ν•΄ 평가 λ˜μ—ˆλ‹€. ν˜„μž¬μ˜ μ—°κ΅¬λŠ” λ”₯ λŸ¬λ‹μ΄ μ•ˆκ³Ό 이미지λ₯Ό 4λ°° 해상도와 ꡬ쑰적인 μ„ΈλΆ€ ν•­λͺ©μ΄ 잘 보이도둝 κ°œμ„ ν•  수 μžˆλ‹€λŠ” 것을 λ³΄μ—¬μ£Όμ—ˆλ‹€. ν–₯μƒλœ μ‹œμ‹ κ²½μœ λ‘ 사진은 μ‹œμ‹ κ²½μ˜ 병리학적인 νŠΉμ„±μ˜ 진단 정확도λ₯Ό λͺ…ν™•νžˆ ν–₯μƒμ‹œν‚¨λ‹€. μ„±λŠ₯평가결과 평균 PSNR은 25.01 SSIM은 0.75 MOSλŠ” 4.33으둜 λ‚˜νƒ€λ‚¬λ‹€. μ„Έλ²ˆμ§Έ μ—°κ΅¬μ—μ„œλŠ” ν™˜μž 정보와 μ•ˆκ³Ό μ˜μƒ(μ‹œμ‹ κ²½μœ λ‘ 사진과 뢉은색이 μ—†λŠ” λ§λ§‰μ‹ κ²½μ„¬μœ μΈ΅ 사진)을 μ΄μš©ν•΄ λ…Ήλ‚΄μž₯ μ˜μ‹¬ ν™˜μžλ₯Ό λΆ„λ³„ν•˜κ³  λ…Ήλ‚΄μž₯ μ˜μ‹¬ ν™˜μžμ˜ λ°œλ³‘ μ—°μˆ˜λ₯Ό μ˜ˆμΈ‘ν•˜λŠ” λ”₯ λŸ¬λ‹ λͺ¨λΈμ„ κ°œλ°œν•˜μ˜€λ‹€. μž„μƒ 데이터듀은 λ…Ήλ‚΄μž₯을 μ§„λ‹¨ν•˜κ±°λ‚˜ μ˜ˆμΈ‘ν•˜λŠ”λ° μœ μš©ν•œ 정보듀을 가지고 μžˆλ‹€. ν•˜μ§€λ§Œ, μ–΄λ–»κ²Œ λ‹€μ–‘ν•œ μœ ν˜•μ˜ μž„μƒμ •λ³΄λ“€μ„ μ‘°ν•©ν•˜λŠ” 것이 각각의 ν™˜μžλ“€μ— λŒ€ν•΄ 잠재적인 λ…Ήλ‚΄μž₯을 μ˜ˆμΈ‘ν•˜λŠ”λ° μ–΄λ–€ 영ν–₯을 μ£ΌλŠ”μ§€μ— λŒ€ν•œ 연ꡬ가 진행 된 적이 μ—†λ‹€. λ…Ήλ‚΄μž₯ 의 μ‹¬μž λΆ„λ₯˜μ™€ λ°œλ³‘ λ…„ 수 μ˜ˆμΈ‘μ€ ν•©μ„± κ³± μžλ™ 인코더(CAE)λ₯Ό λΉ„ 지도적 νŠΉμ„± μΆ”μΆœ 기둜 μ‚¬μš©ν•˜κ³ , κΈ°κ³„ν•™μŠ΅ λΆ„λ₯˜ 기와 νšŒκ·€κΈ°λ₯Ό 톡해 μ§„ν–‰ν•˜μ˜€λ‹€. μ„€κ³„ν•œ λͺ¨λΈμ€ 정확도와 ν‰κ· μ œκ³±μ˜€μ°¨(MSE)λ₯Ό 톡해 평가 λ˜μ—ˆμœΌλ©°, 이미지 νŠΉμ§•κ³Ό ν™˜μž νŠΉμ§•μ€ μ‘°ν•©ν–ˆμ„ λ•Œ λ…Ήλ‚΄μž₯ μ˜μ‹¬ ν™˜μž λΆ„λ₯˜μ™€ λ°œλ³‘ λ…„ 수 예츑의 μ„±λŠ₯이 이미지 νŠΉμ§•κ³Ό ν™˜μž νŠΉμ§•μ„ 각각 썼을 λ•Œλ³΄λ‹€ μ„±λŠ₯이 μ’‹μ•˜λ‹€. μ •λ‹΅κ³Όμ˜ MSEλŠ” 2.613으둜 λ‚˜νƒ€λ‚¬λ‹€. λ³Έ μ—°κ΅¬μ—μ„œλŠ” λ”₯ λŸ¬λ‹μ„ μ΄μš©ν•΄ λ…Ήλ‚΄μž₯ κ΄€λ ¨ μž„μƒ 데이터 쀑 λ§λ§‰μ‹ κ²½μ„¬μœ μΈ΅, μ‹ κ²½μ ˆμ„Έν¬μΈ΅ 사진을 λ…Ήλ‚΄μž₯ 진단에 μ΄μš©λ˜μ—ˆκ³ , μ‹œμ‹ κ²½μœ λ‘ 사진은 μ‹œμ‹ κ²½μ˜ 병리학적인 진단 정확도λ₯Ό λ†’μ˜€κ³ , ν™˜μž μ •λ³΄λŠ” 보닀 μ •ν™•ν•œ λ…Ήλ‚΄μž₯ μ˜μ‹¬ ν™˜μž λΆ„λ₯˜μ™€ λ°œλ³‘ λ…„ 수 μ˜ˆμΈ‘μ— μ΄μš©λ˜μ—ˆλ‹€. ν–₯μƒλœ λ…Ήλ‚΄μž₯ 진단 μ„±λŠ₯은 기술적이고 μž„μƒμ μΈ μ§€ν‘œλ“€μ„ 톡해 κ²€μ¦λ˜μ—ˆλ‹€.This paper presents deep learning-based methods for improving glaucoma diagnosis support systems. Novel methods were applied to glaucoma clinical cases and the results were evaluated. In the first study, a deep learning classifier for glaucoma diagnosis based on spectral-domain optical coherence tomography (SD-OCT) images was proposed and evaluated. Spectral-domain optical coherence tomography (SD-OCT) is commonly employed as an imaging modality for the evaluation of glaucomatous structural damage. The classification model was developed using convolutional neural network (CNN) as a base, and was trained with SD-OCT retinal nerve fiber layer (RNFL) and macular ganglion cell-inner plexiform layer (GCIPL) images. The proposed network architecture, termed Dual-Input Convolutional Neural Network (DICNN), showed great potential as an effective classification algorithm based on two input images. DICNN was trained with both RNFL and GCIPL thickness maps that enabled it to discriminate between normal and glaucomatous eyes. The performance of the proposed DICNN was evaluated with accuracy and area under the receiver operating characteristic curve (AUC), and was compared to other methods using these metrics. Compared to other methods, the proposed DICNN model demonstrated high diagnostic ability for the discrimination of early-stage glaucoma patients in normal subjects. AUC, sensitivity and specificity was 0.869, 0.921, 0.756 respectively. In the second study, a deep-learning method for increasing the resolution and improving the legibility of Optic-disc Photography(ODP) was proposed. ODP has been proven to be useful for optic nerve evaluation in glaucoma. But in clinical practice, limited patient cooperation, small pupil or media opacities can limit the performance of ODP. A model to enhance the resolution of ODP images, termed super-resolution, was developed using Super Resolution Generative Adversarial Network(SR-GAN). To train this model, high-resolution original ODP images were transformed into two counterparts: (1) down-scaled low-resolution ODPs, and (2) compensated high-resolution ODPs with enhanced visibility of the optic disc margin and surrounding retinal vessels which were produced using a customized image post-processing algorithm. The SR-GAN was trained to learn and recognize the differences between these two counterparts. The performance of the network was evaluated using Peak Signal to Noise Ratio (PSNR), Structural Similarity (SSIM), and Mean Opinion Score (MOS). The proposed study demonstrated that deep learning can be applied to create a generative model that is capable of producing enhanced ophthalmic images with 4x resolution and with improved structural details. The proposed method can be used to enhance ODPs and thereby significantly increase the detection accuracy of optic disc pathology. The average PSNR, SSIM and MOS was 25.01, 0.75, 4.33 respectively In the third study, a deep-learning model was used to classify suspected glaucoma and to predict subsequent glaucoma onset-year in glaucoma suspects using clinical data and retinal images (ODP & Red-free Fundus RNFL Photo). Clinical data contains useful information about glaucoma diagnosis and prediction. However, no study has been undertaken to investigate how combining different types of clinical information would be helpful for predicting the subsequent course of glaucoma in an individual patient. For this study, image features extracted using Convolutional Auto Encoder (CAE) along with clinical features were used for glaucoma suspect classification and onset-year prediction. The performance of the proposed model was evaluated using accuracy and Mean Squared Error (MSE). Combing the CAE extracted image features and clinical features improved glaucoma suspect classification and on-set year prediction performance as compared to using the image features and patient features separately. The average MSE between onset-year and predicted onset year was 2.613 In this study, deep learning methodology was applied to clinical images related to glaucoma. DICNN with RNFL and GCIPL images were used for classification of glaucoma, SR-GAN with ODP images were used to increase detection accuracy of optic disc pathology, and CAE & machine learning algorithm with clinical data and retinal images was used for glaucoma suspect classification and onset-year predication. The improved glaucoma diagnosis performance was validated using both technical and clinical parameters. The proposed methods as a whole can significantly improve outcomes of glaucoma patients by early detection, prediction and enhancing detection accuracy.Contents Abstract i Contents iv List of Tables vii List of Figures viii Chapter 1 General Introduction 1 1.1 Glaucoma 1 1.2 Deep Learning for Glaucoma Diagnosis 3 1.4 Thesis Objectives 3 Chapter 2 Dual-Input Convolutional Neural Network for Glaucoma Diagnosis using Spectral-Domain Optical Coherence Tomography 6 2.1 Introduction 6 2.1.1 Background 6 2.1.2 Related Work 7 2.2 Methods 8 2.2.1 Study Design 8 2.2.2 Dataset 9 2.2.3 Dual-Input Convolutional Neural Network (DICNN) 15 2.2.4 Training Environment 18 2.2.5 Statistical Analysis 19 2.3 Results 20 2.3.1 DICNN Performance 20 2.3.1 Grad-CAM for DICNN 34 2.4 Discussion 37 2.4.1 Research Significance 37 2.4.2 Limitations 40 2.5 Conclusion 42 Chapter 3 Deep-learning-based enhanced optic-disc photography 43 3.1 Introduction 43 3.1.1 Background 43 3.1.2 Needs 44 3.1.3 Related Work 45 3.2 Methods 46 3.2.1 Study Design 46 3.2.2 Dataset 46 3.2.2.1 Details on Customized Image Post-Processing Algorithm 47 3.2.3 SR-GAN Network 50 3.2.3.1 Design of Generative Adversarial Network 50 3.2.3.2 Loss Functions 55 3.2.4 Assessment of Clinical Implications of Enhanced ODPs 58 3.2.5 Statistical Analysis 60 3.2.6 Hardware Specifications & Software Specifications 60 3.3 Results 62 3.3.1 Training Loss of Modified SR-GAN 62 3.3.2 Performance of Final Network 66 3.3.3 Clinical Validation of Enhanced ODP by MOS comparison 77 3.3.4 Comparison of DH-Detection Accuracy 79 3.4 Discussion 80 3.4.1 Research Significance 80 3.4.2 Limitations 85 3.5 Conclusion 88 Chapter 4 Deep Learning Based Prediction of Glaucoma Onset Using Retinal Image and Patient Data 89 4.1 Introduction 89 4.1.1 Background 89 4.1.2 Related Work 90 4.2 Methods 90 4.2.1 Study Design 90 4.2.2 Dataset 91 4.2.3 Design of Overall System 94 4.2.4 Design of Convolutional Auto Encoder 95 4.2.5 Glaucoma Suspect Classification 97 4.2.6 Glaucoma Onset-Year Prediction 97 4.3 Result 99 4.3.1 Performance of Designed CAE 99 4.3.2 Performance of Designed Glaucoma Suspect Classification 101 4.3.3 Performance of Designed Glaucoma Onset-Year Prediction 105 4.4 Discussion 110 4.4.1 Research Significance 110 4.4.2 Limitations 110 4.5 Conclusion 111 Chapter 5 Summary and Future Works 112 5.1 Thesis Summary 112 5.2 Limitations and Future Works 113 Bibliography 115 Abstract in Korean 127 Acknowledgement 130Docto

    RFormer: Transformer-based Generative Adversarial Network for Real Fundus Image Restoration on A New Clinical Benchmark

    Full text link
    Ophthalmologists have used fundus images to screen and diagnose eye diseases. However, different equipments and ophthalmologists pose large variations to the quality of fundus images. Low-quality (LQ) degraded fundus images easily lead to uncertainty in clinical screening and generally increase the risk of misdiagnosis. Thus, real fundus image restoration is worth studying. Unfortunately, real clinical benchmark has not been explored for this task so far. In this paper, we investigate the real clinical fundus image restoration problem. Firstly, We establish a clinical dataset, Real Fundus (RF), including 120 low- and high-quality (HQ) image pairs. Then we propose a novel Transformer-based Generative Adversarial Network (RFormer) to restore the real degradation of clinical fundus images. The key component in our network is the Window-based Self-Attention Block (WSAB) which captures non-local self-similarity and long-range dependencies. To produce more visually pleasant results, a Transformer-based discriminator is introduced. Extensive experiments on our clinical benchmark show that the proposed RFormer significantly outperforms the state-of-the-art (SOTA) methods. In addition, experiments of downstream tasks such as vessel segmentation and optic disc/cup detection demonstrate that our proposed RFormer benefits clinical fundus image analysis and applications. The dataset, code, and models are publicly available at https://github.com/dengzhuo-AI/Real-FundusComment: IEEE J-BHI 2022; The First Benchmark and First Transformer-based Method for Real Clinical Fundus Image Restoratio

    GAN-Based Super-Resolution And Segmentation Of Retinal Layers In Optical Coherence Tomography Scans

    Get PDF
    Optical Coherence Tomography (OCT) has been identified as a noninvasive and cost-effective imaging modality for identifying potential biomarkers for Alzheimer\u27s diagnosis and progress detection. Current hypotheses indicate that retinal layer thickness, which can be assessed via OCT scans, is an efficient biomarker for identifying Alzheimer\u27s disease. Due to factors such as speckle noise, a small target region, and unfavorable imaging conditions manual segmentation of retina layers is a challenging task. Therefore, as a reasonable first step, this study focuses on automatically segmenting retinal layers to separate them for subsequent investigations. Another important challenge commonly faced is the lack of clarity of the layer boundaries in retina OCT scans, which compels the research of super-resolving the images for improved clarity. Deep learning pipelines have stimulated substantial progress for the segmentation tasks. Generative adversarial networks (GANs) are a prominent field of deep learning which achieved astonishing performance in semantic segmentation. Conditional adversarial networks as a general-purpose solution to image-to-image translation problems not only learn the mapping from the input image to the output image but also learn a loss function to train this mapping. We propose a GAN-based segmentation model and evaluate incorporating popular networks, namely, U-Net and ResNet, in the GAN architecture with additional blocks of transposed convolution and sub-pixel convolution for the task of upscaling OCT images from low to high resolution by a factor of four. We also incorporate the Dice loss as an additional reconstruction loss term to improve the performance of this joint optimization task. Our best model configuration empirically achieved the Dice coefficient of 0.867 and mIOU of 0.765

    Enhancing Image Quality: A Comparative Study of Spatial, Frequency Domain, and Deep Learning Methods

    Get PDF
    Image restoration and noise reduction methods have been created to restore deteriorated images and improve their quality. These methods have garnered substantial significance in recent times, mainly due to the growing utilization of digital imaging across diverse domains, including but not limited to medical imaging, surveillance, satellite imaging, and numerous others. In this paper, we conduct a comparative analysis of three distinct approaches to image restoration: the spatial method, the frequency domain method, and the deep learning method. The study was conducted on a dataset of 10,000 images, and the performance of each method was evaluated using the accuracy and loss metrics. The results show that the deep learning method outperformed the other two methods, achieving a validation accuracy of 72.68% after 10 epochs. The spatial method had the lowest accuracy of the three, achieving a validation accuracy of 69.98% after 10 epochs. The FFT frequency domain method had a validation accuracy of 52.87% after 10 epochs, significantly lower than the other two methods. The study demonstrates that deep learning is a promising approach for image classification tasks and outperforms traditional methods such as spatial and frequency domain techniques
    • …
    corecore