2,192 research outputs found
An Interval Type-2 Fuzzy Approach to Automatic PDF Generation for Histogram Specification
Image enhancement plays an important role in several application in the field
of computer vision and image processing. Histogram specification (HS) is one of
the most widely used techniques for contrast enhancement of an image, which
requires an appropriate probability density function for the transformation. In
this paper, we propose a fuzzy method to find a suitable PDF automatically for
histogram specification using interval type - 2 (IT2) fuzzy approach, based on
the fuzzy membership values obtained from the histogram of input image. The
proposed algorithm works in 5 stages which includes - symmetric Gaussian
fitting on the histogram, extraction of IT2 fuzzy membership functions (MFs)
and therefore, footprint of uncertainty (FOU), obtaining membership value (MV),
generating PDF and application of HS. We have proposed 4 different methods to
find membership values - point-wise method, center of weight method, area
method, and karnik-mendel (KM) method. The framework is sensitive to local
variations in the histogram and chooses the best PDF so as to improve contrast
enhancement. Experimental validity of the methods used is illustrated by
qualitative and quantitative analysis on several images using the image quality
index - Average Information Content (AIC) or Entropy, and by comparison with
the commonly used algorithms such as Histogram Equalization (HE), Recursive
Mean-Separate Histogram Equalization (RMSHE) and Brightness Preserving Fuzzy
Histogram Equalization (BPFHE). It has been found out that on an average, our
algorithm improves the AIC index by 11.5% as compared to the index obtained by
histogram equalisation
Bio-Authentication based Secure Transmission System using Steganography
Biometrics deals with identity verification of an individual by using certain
physiological or behavioral features associated with a person. Biometric
identification systems using fingerprints patterns are called AFIS (Automatic
Fingerprint Identification System). In this paper a composite method for
Fingerprint recognition is considered using a combination of Fast Fourier
Transform (FFT) and Sobel Filters for improvement of a poor quality fingerprint
image. Steganography hides messages inside other messages in such a way that an
"adversary" would not even know a secret message were present. The objective of
our paper is to make a bio-secure system. In this paper bio-authentication has
been implemented in terms of finger print recognition and the second part of
the paper is an interactive steganographic system hides the user's data by two
options- creating a songs list or hiding the data in an image.Comment: IEEE Publication format, International Journal of Computer Science
and Information Security, IJCSIS, Vol. 8 No. 1, April 2010, USA. ISSN 1947
5500, http://sites.google.com/site/ijcsis
Enhancing the Accuracy of Biometric Feature Extraction Fusion Using Gabor Filter and Mahalanobis Distance Algorithm
Biometric recognition systems have advanced significantly in the last decade
and their use in specific applications will increase in the near future. The
ability to conduct meaningful comparisons and assessments will be crucial to
successful deployment and increasing biometric adoption. The best modality used
as unimodal biometric systems are unable to fully address the problem of higher
recognition rate. Multimodal biometric systems are able to mitigate some of the
limitations encountered in unimodal biometric systems, such as
non-universality, distinctiveness, non-acceptability, noisy sensor data, spoof
attacks, and performance. More reliable recognition accuracy and performance
are achievable as different modalities were being combined together and
different algorithms or techniques were being used. The work presented in this
paper focuses on a bimodal biometric system using face and fingerprint. An
image enhancement technique (histogram equalization) is used to enhance the
face and fingerprint images. Salient features of the face and fingerprint were
extracted using the Gabor filter technique. A dimensionality reduction
technique was carried out on both images extracted features using a principal
component analysis technique. A feature level fusion algorithm (Mahalanobis
distance technique) is used to combine each unimodal feature together. The
performance of the proposed approach is validated and is effective.Comment: Focused on extraction of feature from two different modalities (face
and fingerprint) using Gabor filte
A Bio-Inspired Multi-Exposure Fusion Framework for Low-light Image Enhancement
Low-light images are not conducive to human observation and computer vision
algorithms due to their low visibility. Although many image enhancement
techniques have been proposed to solve this problem, existing methods
inevitably introduce contrast under- and over-enhancement. Inspired by human
visual system, we design a multi-exposure fusion framework for low-light image
enhancement. Based on the framework, we propose a dual-exposure fusion
algorithm to provide an accurate contrast and lightness enhancement.
Specifically, we first design the weight matrix for image fusion using
illumination estimation techniques. Then we introduce our camera response model
to synthesize multi-exposure images. Next, we find the best exposure ratio so
that the synthetic image is well-exposed in the regions where the original
image is under-exposed. Finally, the enhanced result is obtained by fusing the
input image and the synthetic image according to the weight matrix. Experiments
show that our method can obtain results with less contrast and lightness
distortion compared to that of several state-of-the-art methods.Comment: Project website: https://baidut.github.io/BIMEF
Comparative analysis of evolutionary algorithms for image enhancement
Evolutionary algorithms are metaheuristic techniques that derive inspiration
from the natural process of evolution. They can efficiently solve (generate
acceptable quality of solution in reasonable time) complex optimization
(NP-Hard) problems. In this paper, automatic image enhancement is considered as
an optimization problem and three evolutionary algorithms (Genetic Algorithm,
Differential Evolution and Self Organizing Migration Algorithm) are employed to
search for an optimum solution. They are used to find an optimum parameter set
for an image enhancement transfer function. The aim is to maximize a fitness
criterion which is a measure of image contrast and the visibility of details in
the enhanced image. The enhancement results obtained using all three
evolutionary algorithms are compared amongst themselves and also with the
output of histogram equalization method
An Image Based Technique for Enhancement of Underwater Images
The underwater images usually suffers from non-uniform lighting, low
contrast, blur and diminished colors. In this paper, we proposed an image based
preprocessing technique to enhance the quality of the underwater images. The
proposed technique comprises a combination of four filters such as homomorphic
filtering, wavelet denoising, bilateral filter and contrast equalization. These
filters are applied sequentially on degraded underwater images. The literature
survey reveals that image based preprocessing algorithms uses standard filter
techniques with various combinations. For smoothing the image, the image based
preprocessing algorithms uses the anisotropic filter. The main drawback of the
anisotropic filter is that iterative in nature and computation time is high
compared to bilateral filter. In the proposed technique, in addition to other
three filters, we employ a bilateral filter for smoothing the image. The
experimentation is carried out in two stages. In the first stage, we have
conducted various experiments on captured images and estimated optimal
parameters for bilateral filter. Similarly, optimal filter bank and optimal
wavelet shrinkage function are estimated for wavelet denoising. In the second
stage, we conducted the experiments using estimated optimal parameters, optimal
filter bank and optimal wavelet shrinkage function for evaluating the proposed
technique. We evaluated the technique using quantitative based criteria such as
a gradient magnitude histogram and Peak Signal to Noise Ratio (PSNR). Further,
the results are qualitatively evaluated based on edge detection results. The
proposed technique enhances the quality of the underwater images and can be
employed prior to apply computer vision techniques
A cascade network for Detecting COVID-19 using chest x-rays
The worldwide spread of pneumonia caused by a novel coronavirus poses an
unprecedented challenge to the world's medical resources and prevention and
control measures. Covid-19 attacks not only the lungs, making it difficult to
breathe and life-threatening, but also the heart, kidneys, brain and other
vital organs of the body, with possible sequela. At present, the detection of
COVID-19 needs to be realized by the reverse transcription-polymerase Chain
Reaction (RT-PCR). However, many countries are in the outbreak period of the
epidemic, and the medical resources are very limited. They cannot provide
sufficient numbers of gene sequence detection, and many patients may not be
isolated and treated in time. Given this situation, we researched the
analytical and diagnostic capabilities of deep learning on chest radiographs
and proposed Cascade-SEMEnet which is cascaded with SEME-ResNet50 and
SEME-DenseNet169. The two cascade networks of Cascade - SEMEnet both adopt
large input sizes and SE-Structure and use MoEx and histogram equalization to
enhance the data. We first used SEME-ResNet50 to screen chest X-ray and
diagnosed three classes: normal, bacterial, and viral pneumonia. Then we used
SEME-DenseNet169 for fine-grained classification of viral pneumonia and
determined if it is caused by COVID-19. To exclude the influence of
non-pathological features on the network, we preprocessed the data with U-Net
during the training of SEME-DenseNet169. The results showed that our network
achieved an accuracy of 85.6\% in determining the type of pneumonia infection
and 97.1\% in the fine-grained classification of COVID-19. We used Grad-CAM to
visualize the judgment based on the model and help doctors understand the chest
radiograph while verifying the effectivene
Design of Novel Algorithm and Architecture for Gaussian Based Color Image Enhancement System for Real Time Applications
This paper presents the development of a new algorithm for Gaussian based
color image enhancement system. The algorithm has been designed into
architecture suitable for FPGA/ASIC implementation. The color image enhancement
is achieved by first convolving an original image with a Gaussian kernel since
Gaussian distribution is a point spread function which smoothen the image.
Further, logarithm-domain processing and gain/offset corrections are employed
in order to enhance and translate pixels into the display range of 0 to 255.
The proposed algorithm not only provides better dynamic range compression and
color rendition effect but also achieves color constancy in an image. The
design exploits high degrees of pipelining and parallel processing to achieve
real time performance. The design has been realized by RTL compliant Verilog
coding and fits into a single FPGA with a gate count utilization of 321,804.
The proposed method is implemented using Xilinx Virtex-II Pro XC2VP40-7FF1148
FPGA device and is capable of processing high resolution color motion pictures
of sizes of up to 1600x1200 pixels at the real time video rate of 116 frames
per second. This shows that the proposed design would work for not only still
images but also for high resolution video sequences.Comment: 15 pages, 15 figure
Auto-Classification of Retinal Diseases in the Limit of Sparse Data Using a Two-Streams Machine Learning Model
Automatic clinical diagnosis of retinal diseases has emerged as a promising
approach to facilitate discovery in areas with limited access to specialists.
Based on the fact that fundus structure and vascular disorders are the main
characteristics of retinal diseases, we propose a novel visual-assisted
diagnosis hybrid model mixing the support vector machine (SVM) and deep neural
networks (DNNs). Furthermore, we present a new clinical retina dataset, called
EyeNet2, for ophthalmology incorporating 52 retina diseases classes. Using
EyeNet2, our model achieves 90.43\% diagnosis accuracy, and the model
performance is comparable to the professional ophthalmologists.Comment: A extension work of a workshop paper arXiv admin note: substantial
text overlap with arXiv:1806.0642
Generation of High Dynamic Range Illumination from a Single Image for the Enhancement of Undesirably Illuminated Images
This paper presents an algorithm that enhances undesirably illuminated images
by generating and fusing multi-level illuminations from a single image.The
input image is first decomposed into illumination and reflectance components by
using an edge-preserving smoothing filter. Then the reflectance component is
scaled up to improve the image details in bright areas. The illumination
component is scaled up and down to generate several illumination images that
correspond to certain camera exposure values different from the original. The
virtual multi-exposure illuminations are blended into an enhanced illumination,
where we also propose a method to generate appropriate weight maps for the tone
fusion. Finally, an enhanced image is obtained by multiplying the equalized
illumination and enhanced reflectance. Experiments show that the proposed
algorithm produces visually pleasing output and also yields comparable
objective results to the conventional enhancement methods, while requiring
modest computational loads
- …