867 research outputs found

    Machine Learning And Image Processing For Noise Removal And Robust Edge Detection In The Presence Of Mixed Noise

    Get PDF
    The central goal of this dissertation is to design and model a smoothing filter based on the random single and mixed noise distribution that would attenuate the effect of noise while preserving edge details. Only then could robust, integrated and resilient edge detection methods be deployed to overcome the ubiquitous presence of random noise in images. Random noise effects are modeled as those that could emanate from impulse noise, Gaussian noise and speckle noise. In the first step, evaluation of methods is performed based on an exhaustive review on the different types of denoising methods which focus on impulse noise, Gaussian noise and their related denoising filters. These include spatial filters (linear, non-linear and a combination of them), transform domain filters, neural network-based filters, numerical-based filters, fuzzy based filters, morphological filters, statistical filters, and supervised learning-based filters. In the second step, switching adaptive median and fixed weighted mean filter (SAMFWMF) which is a combination of linear and non-linear filters, is introduced in order to detect and remove impulse noise. Then, a robust edge detection method is applied which relies on an integrated process including non-maximum suppression, maximum sequence, thresholding and morphological operations. The results are obtained on MRI and natural images. In the third step, a combination of transform domain-based filter which is a combination of dual tree – complex wavelet transform (DT-CWT) and total variation, is introduced in order to detect and remove Gaussian noise as well as mixed Gaussian and Speckle noise. Then, a robust edge detection is applied in order to track the true edges. The results are obtained on medical ultrasound and natural images. In the fourth step, a smoothing filter, which is a feed-forward convolutional network (CNN) is introduced to assume a deep architecture, and supported through a specific learning algorithm, l2 loss function minimization, a regularization method, and batch normalization all integrated in order to detect and remove impulse noise as well as mixed impulse and Gaussian noise. Then, a robust edge detection is applied in order to track the true edges. The results are obtained on natural images for both specific and non-specific noise-level

    Color Image Processing based on Graph Theory

    Full text link
    [ES] La visión artificial es uno de los campos en mayor crecimiento en la actualidad que, junto con otras tecnologías como la Biometría o el Big Data, se ha convertido en el foco de interés de numerosas investigaciones y es considerada como una de las tecnologías del futuro. Este amplio campo abarca diversos métodos entre los que se encuentra el procesamiento y análisis de imágenes digitales. El éxito del análisis de imágenes y otras tareas de procesamiento de alto nivel, como pueden ser el reconocimiento de patrones o la visión 3D, dependerá en gran medida de la buena calidad de las imágenes de partida. Hoy en día existen multitud de factores que dañan las imágenes dificultando la obtención de imágenes de calidad óptima, esto ha convertido el (pre-) procesamiento digital de imágenes en un paso fundamental previo a la aplicación de cualquier otra tarea de procesado. Los factores más comunes son el ruido y las malas condiciones de adquisición: los artefactos provocados por el ruido dificultan la interpretación adecuada de la imagen y la adquisición en condiciones de iluminación o exposición deficientes, como escenas dinámicas, causan pérdida de información de la imagen que puede ser clave para ciertas tareas de procesamiento. Los pasos de (pre-)procesamiento de imágenes conocidos como suavizado y realce se aplican comúnmente para solventar estos problemas: El suavizado tiene por objeto reducir el ruido mientras que el realce se centra en mejorar o recuperar la información imprecisa o dañada. Con estos métodos conseguimos reparar información de los detalles y bordes de la imagen con una nitidez insuficiente o un contenido borroso que impide el (post-)procesamiento óptimo de la imagen. Existen numerosos métodos que suavizan el ruido de una imagen, sin embargo, en muchos casos el proceso de filtrado provoca emborronamiento en los bordes y detalles de la imagen. De igual manera podemos encontrar una enorme cantidad de técnicas de realce que intentan combatir las pérdidas de información, sin embargo, estas técnicas no contemplan la existencia de ruido en la imagen que procesan: ante una imagen ruidosa, cualquier técnica de realce provocará también un aumento del ruido. Aunque la idea intuitiva para solucionar este último caso será el previo filtrado y posterior realce, este enfoque ha demostrado no ser óptimo: el filtrado podrá eliminar información que, a su vez, podría no ser recuperable en el siguiente paso de realce. En la presente tesis doctoral se propone un modelo basado en teoría de grafos para el procesamiento de imágenes en color. En este modelo, se construye un grafo para cada píxel de tal manera que sus propiedades permiten caracterizar y clasificar dicho pixel. Como veremos, el modelo propuesto es robusto y capaz de adaptarse a una gran variedad de aplicaciones. En particular, aplicamos el modelo para crear nuevas soluciones a los dos problemas fundamentales del procesamiento de imágenes: suavizado y realce. Se ha estudiado el modelo en profundidad en función del umbral, parámetro clave que asegura la correcta clasificación de los píxeles de la imagen. Además, también se han estudiado las posibles características y posibilidades del modelo que nos han permitido sacarle el máximo partido en cada una de las posibles aplicaciones. Basado en este modelo se ha diseñado un filtro adaptativo capaz de eliminar ruido gaussiano de una imagen sin difuminar los bordes ni perder información de los detalles. Además, también ha permitido desarrollar un método capaz de realzar los bordes y detalles de una imagen al mismo tiempo que se suaviza el ruido presente en la misma. Esta aplicación simultánea consigue combinar dos operaciones opuestas por definición y superar así los inconvenientes presentados por el enfoque en dos etapas.[CA] La visió artificial és un dels camps en major creixement en l'actualitat que, junt amb altres tecnlogies com la Biometria o el Big Data, s'ha convertit en el focus d'interés de nombroses investigacions i és considerada com una de les tecnologies del futur. Aquest ampli camp comprén diversos m`etodes entre els quals es troba el processament digital d'imatges i anàlisis d'imatges digitals. L'èxit de l'anàlisis d'imatges i altres tasques de processament d'alt nivell, com poden ser el reconeixement de patrons o la visió 3D, dependrà en gran manera de la bona qualitat de les imatges de partida. Avui dia existeixen multitud de factors que danyen les imatges dificultant l'obtenció d'imatges de qualitat òptima, açò ha convertit el (pre-) processament digital d'imatges en un pas fonamental previa la l'aplicació de qualsevol altra tasca de processament. Els factors més comuns són el soroll i les males condicions d'adquisició: els artefactes provocats pel soroll dificulten la inter- pretació adequada de la imatge i l'adquisició en condicions d'il·luminació o exposició deficients, com a escenes dinàmiques, causen pèrdua d'informació de la imatge que pot ser clau per a certes tasques de processament. Els passos de (pre-) processament d'imatges coneguts com suavitzat i realç s'apliquen comunament per a resoldre aquests problemes: El suavitzat té com a objecte reduir el soroll mentres que el real se centra a millorar o recuperar la informació imprecisa o danyada. Amb aquests mètodes aconseguim reparar informació dels detalls i bords de la imatge amb una nitidesa insuficient o un contingut borrós que impedeix el (post-)processament òptim de la imatge. Existeixen nombrosos mètodes que suavitzen el soroll d'una imatge, no obstant això, en molts casos el procés de filtrat provoca emborronamiento en els bords i detalls de la imatge. De la mateixa manera podem trobar una enorme quantitat de tècniques de realç que intenten combatre les pèrdues d'informació, no obstant això, aquestes tècniques no contemplen l'existència de soroll en la imatge que processen: davant d'una image sorollosa, qualsevol tècnica de realç provocarà també un augment del soroll. Encara que la idea intuïtiva per a solucionar aquest últim cas seria el previ filtrat i posterior realç, aquest enfocament ha demostrat no ser òptim: el filtrat podria eliminar informació que, al seu torn, podria no ser recuperable en el seguënt pas de realç. En la present Tesi doctoral es proposa un model basat en teoria de grafs per al processament d'imatges en color. En aquest model, es construïx un graf per a cada píxel de tal manera que les seues propietats permeten caracteritzar i classificar el píxel en quëstió. Com veurem, el model proposat és robust i capaç d'adaptar-se a una gran varietat d'aplicacions. En particular, apliquem el model per a crear noves solucions als dos problemes fonamentals del processament d'imatges: suavitzat i realç. S'ha estudiat el model en profunditat en funció del llindar, paràmetre clau que assegura la correcta classificació dels píxels de la imatge. A més, també s'han estudiat les possibles característiques i possibilitats del model que ens han permés traure-li el màxim partit en cadascuna de les possibles aplicacions. Basat en aquest model s'ha dissenyat un filtre adaptatiu capaç d'eliminar soroll gaussià d'una imatge sense difuminar els bords ni perdre informació dels detalls. A més, també ha permés desenvolupar un mètode capaç de realçar els bords i detalls d'una imatge al mateix temps que se suavitza el soroll present en la mateixa. Aquesta aplicació simultània aconseguix combinar dues operacions oposades per definició i superar així els inconvenients presentats per l'enfocament en dues etapes.[EN] Computer vision is one of the fastest growing fields at present which, along with other technologies such as Biometrics or Big Data, has become the focus of interest of many research projects and it is considered one of the technologies of the future. This broad field includes a plethora of digital image processing and analysis tasks. To guarantee the success of image analysis and other high-level processing tasks as 3D imaging or pattern recognition, it is critical to improve the quality of the raw images acquired. Nowadays all images are affected by different factors that hinder the achievement of optimal image quality, making digital image processing a fundamental step prior to the application of any other practical application. The most common of these factors are noise and poor acquisition conditions: noise artefacts hamper proper image interpretation of the image; and acquisition in poor lighting or exposure conditions, such as dynamic scenes, causes loss of image information that can be key for certain processing tasks. Image (pre-) processing steps known as smoothing and sharpening are commonly applied to overcome these inconveniences: Smoothing is aimed at reducing noise and sharpening at improving or recovering imprecise or damaged information of image details and edges with insufficient sharpness or blurred content that prevents optimal image (post-)processing. There are many methods for smoothing the noise in an image, however in many cases the filtering process causes blurring at the edges and details of the image. Besides, there are also many sharpening techniques, which try to combat the loss of information due to blurring of image texture and need to contemplate the existence of noise in the image they process. When dealing with a noisy image, any sharpening technique may amplify the noise. Although the intuitive idea to solve this last case would be the previous filtering and later sharpening, this approach has proved not to be optimal: the filtering could remove information that, in turn, may not be recoverable in the later sharpening step. In the present PhD dissertation we propose a model based on graph theory for color image processing from a vector approach. In this model, a graph is built for each pixel in such a way that its features allow to characterize and classify the pixel. As we will show, the model we proposed is robust and versatile: potentially able to adapt to a variety of applications. In particular, we apply the model to create new solutions for the two fundamentals problems in image processing: smoothing and sharpening. To approach high performance image smoothing we use the proposed model to determine if a pixel belongs to a at region or not, taking into account the need to achieve a high-precision classification even in the presence of noise. Thus, we build an adaptive soft-switching filter by employing the pixel classification to combine the outputs from a filter with high smoothing capability and a softer one to smooth edge/detail regions. Further, another application of our model allows to use pixels characterization to successfully perform a simultaneous smoothing and sharpening of color images. In this way, we address one of the classical challenges within the image processing field. We compare all the image processing techniques proposed with other state-of-the-art methods to show that they are competitive both from an objective (numerical) and visual evaluation point of view.Pérez Benito, C. (2019). Color Image Processing based on Graph Theory [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/123955TESI

    Improved methods for finger vein identification using composite median-wiener filter and hierarchical centroid features extraction

    Get PDF
    Finger vein identification is a potential new area in biometric systems. Finger vein patterns contain highly discriminative characteristics, which are difficult to be forged because they reside underneath the skin of the finger and require a specific device to capture them. Research have been carried out in this field but there is still an unresolved issue related to low-quality data due to data capturing and processing. Low-quality data have caused errors in the feature extraction process and reduced identification performance rate in finger vein identification. To address this issue, a new image enhancement and feature extraction methods were developed to improve finger vein identification. The image enhancement, Composite Median-Wiener (CMW) filter would improve image quality and preserve the edges of the finger vein image. Next, the feature extraction method, Hierarchical Centroid Feature Method (HCM) was fused with statistical pixel-based distribution feature method at the feature-level fusion to improve the performance of finger vein identification. These methods were evaluated on public SDUMLA-HMT and FV-USM finger vein databases. Each database was divided into training and testing sets. The average result of the experiments conducted was taken to ensure the accuracy of the measurements. The k-Nearest Neighbor classifier with city block distance to match the features was implemented. Both these methods produced accuracy as high as 97.64% for identification rate and 1.11% of equal error rate (EER) for measures verification rate. These showed that the accuracy of the proposed finger vein identification method is higher than the one reported in the literature. As a conclusion, the results have proven that the CMW filter and HCM have significantly improved the accuracy of finger vein identification

    A Comprehensive Review of Image Restoration and Noise Reduction Techniques

    Get PDF
    Images play a crucial role in modern life and find applications in diverse fields, ranging from preserving memories to conducting scientific research. However, images often suffer from various forms of degradation such as blur, noise, and contrast loss. These degradations make images difficult to interpret, reduce their visual quality, and limit their practical applications. To overcome these challenges, image restoration and noise reduction techniques have been developed to recover degraded images and enhance their quality. These techniques have gained significant importance in recent years, especially with the increasing use of digital imaging in various fields such as medical imaging, surveillance, satellite imaging, and many others. This paper presents a comprehensive review of image restoration and noise reduction techniques, encompassing spatial and frequency domain methods, and deep learning-based techniques. The paper also discusses the evaluation metrics utilized to assess the effectiveness of these techniques and explores future research directions in this field. The primary objective of this paper is to offer a comprehensive understanding of the concepts and methods involved in image restoration and noise reduction

    FINGERPRINT ENHANCEMENT USING FUZZY LOGIC AND DEEP NEURAL NETWORKS

    Get PDF
    Department of Computer Science and EngineeringFingerprint recognition analysis is one of the most leading preferred prodigious biometric advancement which has drawn generous consideration in biometrics. In this work, fingerprint Intensification is performed which is defined by Fuzzy logic technique and recognize the matching image with its unique characteristics extracted and classify the features extracted from a fuzzy enhanced image along with three major types of Neural Networks which are Feed Forward Artificial Neural Network, Neural Network, Recurrent Neural Network in order to classify the unique features extracted from a fingerprint image. This work efficiently expresses the results with Fuzzy logic enhancement and Neural Networks classifiers. Its principle goal is to improve the image using Fuzzy and extricate the spurious minutiae detected and classify the different features generated using GLCM and DWT. This work displays a framework of unique finger impression classification based on particular characteristics for extricating different features and three types of Neural Network for classification. Fuzzy technique is used for the fuzzy based image enhancement to urge the clear see of the unique finger impression. Fingerprint Image Intensification is the procedure to enhance the distorted images to encourage the recognizable proof. The motivation behind the work is to enrich the quality of the distorted condition image generated from any fingerprint sensor, as Images can be corrupted due to various conditions and one of the principal issues is the resolution of the fingerprint sensor generating noisy images. High-quality pictures are vital for the exact coordinating of unique finger impression pictures. But unique mark pictures are seldom of idealizing refinement. As it may be corrupted or debased due to varieties of the skin, impression state and condition. In this way, unique finger impression images must be improved before utilized. The idea behind this work fingerprint image intensification process is to improve the quality of distorted and noisy fingerprint images generated from a low-cost fingerprint sensor. Execution of current ???ngerprint acknowledgment frameworks is vigorously in???uenced by the precision of their characteristic???s extraction evaluation. These days, there are more ways to deal with ???ngerprint analysis with worthy outcomes. Issues begin to emerge in low-quality conditions where the dominant part of the conventional strategies dependent on examining the surface of ???ngerprint can't handle this issue so e???ectively as Neural Networks. Fuzzy logic technique is implemented first to remediate the distorted picture and enhance it with the implementation of GLCM and DWT2 algorithm features of an image is extracted, post to which three types of Neural Network Classification is performed to analyze the accuracy of the image generated from the extracted feature parameters and match the test and trained result with the implementation of Neural Networks and classify the outcome results. The three Neural Network used is Artificial Neural Network (ANN), Neural Network (NN), Recurrent Neural Network (RNN). This algorithm works efficiently to identify the fingerprint matching from the predefined trained images from the fuzzy enhanced image generated. Experiments are performed (in MATLAB 2019 student version) to make sure the extraction process should not get the false minutiae and preserve the true extracted features Fuzzy based Image Enhancement method makes sure the feature traits of the image is intensified. Better improvement proves the quality improvement further incrementing the highest accuracy determined in the classification further. This work can be used in a wide area of applications in biometrics as it is a combined work of distorted fingerprints enhancement, false feature removal, true feature extraction, matching of the images for identification purpose and classification using Neural Networks. Experiments show results which are quite promising and gives a direction of the subsequent further analysis in future work.clos

    Adaptive Methods for Point Cloud and Mesh Processing

    Get PDF
    Point clouds and 3D meshes are widely used in numerous applications ranging from games to virtual reality to autonomous vehicles. This dissertation proposes several approaches for noise removal and calibration of noisy point cloud data and 3D mesh sharpening methods. Order statistic filters have been proven to be very successful in image processing and other domains as well. Different variations of order statistics filters originally proposed for image processing are extended to point cloud filtering in this dissertation. A brand-new adaptive vector median is proposed in this dissertation for removing noise and outliers from noisy point cloud data. The major contributions of this research lie in four aspects: 1) Four order statistic algorithms are extended, and one adaptive filtering method is proposed for the noisy point cloud with improved results such as preserving significant features. These methods are applied to standard models as well as synthetic models, and real scenes, 2) A hardware acceleration of the proposed method using Microsoft parallel pattern library for filtering point clouds is implemented using multicore processors, 3) A new method for aerial LIDAR data filtering is proposed. The objective is to develop a method to enable automatic extraction of ground points from aerial LIDAR data with minimal human intervention, and 4) A novel method for mesh color sharpening using the discrete Laplace-Beltrami operator is proposed. Median and order statistics-based filters are widely used in signal processing and image processing because they can easily remove outlier noise and preserve important features. This dissertation demonstrates a wide range of results with median filter, vector median filter, fuzzy vector median filter, adaptive mean, adaptive median, and adaptive vector median filter on point cloud data. The experiments show that large-scale noise is removed while preserving important features of the point cloud with reasonable computation time. Quantitative criteria (e.g., complexity, Hausdorff distance, and the root mean squared error (RMSE)), as well as qualitative criteria (e.g., the perceived visual quality of the processed point cloud), are employed to assess the performance of the filters in various cases corrupted by different noisy models. The adaptive vector median is further optimized for denoising or ground filtering aerial LIDAR data point cloud. The adaptive vector median is also accelerated on multi-core CPUs using Microsoft Parallel Patterns Library. In addition, this dissertation presents a new method for mesh color sharpening using the discrete Laplace-Beltrami operator, which is an approximation of second order derivatives on irregular 3D meshes. The one-ring neighborhood is utilized to compute the Laplace-Beltrami operator. The color for each vertex is updated by adding the Laplace-Beltrami operator of the vertex color weighted by a factor to its original value. Different discretizations of the Laplace-Beltrami operator have been proposed for geometrical processing of 3D meshes. This work utilizes several discretizations of the Laplace-Beltrami operator for sharpening 3D mesh colors and compares their performance. Experimental results demonstrated the effectiveness of the proposed algorithms

    Fuzzy machine vision based inspection

    Get PDF
    Machine vision system has been fostered to solve many realistic problems in various fields. Its role in achieving superior quality and productivity is of paramount importance. But, for such system to be attractive, it needs to be fast, accurate and cost-effective. This dissertation is based on a number of practical machine vision based inspection projects obtained from the automotive industry. It presents a collection of developed efficient fuzzy machine vision approaches endorsed with experimental results. It also covers the conceptual design, development and testing of various fuzzy machine vision based inspection approaches for different industrial applications. To assist in developing and evaluating the performance of the proposed approaches, several parts are tested under varying lighting conditions. This research deals with two important aspects of machine vision based inspection. In the first part, it concentrates on the topics of component detection and component orientation identification. The components used in this part are metal clips mounted on a dash panel frame that is installed in the door of trucks. Therefore, we propose a fuzzy machine vision based clip detection model and a fuzzy machine vision based clip orientation identification model to inspect the proper placement of clips on dash panels. Both models are efficient and fast in terms of accuracy and processing time. In the second part of the research, we are dealing with machined part defects such as broken edge, porosity and tool marks. The se defects occur on the surface of die cast aluminum automotive pump housings. As a result, an automated fuzzy machine vision based broken edge detection method, an efficient fuzzy machine vision based porosity detection technique and a neuro-fuzzy part classification model based on tool marks are developed. Computational results show that the proposed approaches are effective in yielding satisfactory results to the tested image databases. There are four main contributions to this work. The first contribution is the development of the concept of composite matrices in conjunction with XOR feature extractor using fuzzy subtractive clustering for clip detection. The second contribution is about a proposed model based on grouping and counting pixels in pre-selective areas which tracks pixel colors in separated RGB channels to determine whether the orientation of the clip is acceptable or not. The construction of three novel edge based features embedded in fuzzy C-means clustering for broken edge detection marks the third contribution. At last, the fourth contribution presents the core of porosity candidates concept and its correlation with twelve developed matrices. This, in turn, results in the development of five different features used in our fuzzy machine vision based porosity detection approach
    corecore