624 research outputs found
Wavelet Integrated CNNs for Noise-Robust Image Classification
Convolutional Neural Networks (CNNs) are generally prone to noise
interruptions, i.e., small image noise can cause drastic changes in the output.
To suppress the noise effect to the final predication, we enhance CNNs by
replacing max-pooling, strided-convolution, and average-pooling with Discrete
Wavelet Transform (DWT). We present general DWT and Inverse DWT (IDWT) layers
applicable to various wavelets like Haar, Daubechies, and Cohen, etc., and
design wavelet integrated CNNs (WaveCNets) using these layers for image
classification. In WaveCNets, feature maps are decomposed into the
low-frequency and high-frequency components during the down-sampling. The
low-frequency component stores main information including the basic object
structures, which is transmitted into the subsequent layers to extract robust
high-level features. The high-frequency components, containing most of the data
noise, are dropped during inference to improve the noise-robustness of the
WaveCNets. Our experimental results on ImageNet and ImageNet-C (the noisy
version of ImageNet) show that WaveCNets, the wavelet integrated versions of
VGG, ResNets, and DenseNet, achieve higher accuracy and better noise-robustness
than their vanilla versions.Comment: CVPR accepted pape
Detection and estimation of image blur
The airborne imagery consisting of infrared (IR) and multispectral (MSI) images collected in 2009 under airborne mine and minefield detection program by Night Vision and Electronic Sensors Directorate (NVESD) was found to be severely blurred due to relative motion between the camera and the object and some of them with defocus blurs due to various reasons. Automated detection of blur due to motion and defocus blurs and the estimation of blur like point spread function for severely degraded images is an important task for processing and detection in such airborne imagery. Although several full reference and reduced reference methods are available in the literature, using no reference methods are desirable because there was no information of the degradation function and the original image data. In this thesis, three no reference algorithms viz. Haar wavelet (HAAR), modified Haar using singular value decomposition (SVD), and intentional blurring pixel difference (IBD) for blur detection are compared and their performance is qualified based on missed detections and false alarms. Three human subjects were chosen to perform subjective testing on randomly selected data sets and the truth for each frame was obtained from majority voting. The modified Haar algorithm (SVD) resulted in the least number of missed detections and least number of false alarms. This thesis also evaluates several methods for estimating the point spread function (PSF) of these degraded images. The Auto-correlation function (ACF), Hough transform (Hough) and steer Gaussian filter (SGF) based methods were tested on several synthetically motion blurred images and further validated on naturally blurred images. Statistics of pixel error estimate using these methods were computed based on 8640 artificially blurred image frames --Abstract, page iii
Techniques for enhancing digital images
The images obtain from either research studies or optical instruments are
often corrupted with noise. Image denoising involves the manipulation of image
data to produce a visually high quality image. This thesis reviews the existing
denoising algorithms and the filtering approaches available for enhancing images
and/or data transmission.
Spatial-domain and Transform-domain digital image filtering algorithms
have been used in the past to suppress different noise models. The different noise
models can be either additive or multiplicative. Selection of the denoising algorithm
is application dependent. It is necessary to have knowledge about the noise present
in the image so as to select the appropriated denoising algorithm. Noise models
may include Gaussian noise, Salt and Pepper noise, Speckle noise and Brownian
noise. The Wavelet Transform is similar to the Fourier transform with a completely
different merit function. The main difference between Wavelet transform and
Fourier transform is that, in the Wavelet Transform, Wavelets are localized in both
time and frequency. In the standard Fourier Transform, Wavelets are only localized
in frequency. Wavelet analysis consists of breaking up the signal into shifted and
scales versions of the original (or mother) Wavelet. The Wiener Filter (mean
squared estimation error) finds implementations as a LMS filter (least mean
squares), RLS filter (recursive least squares), or Kalman filter.
Quantitative measure (metrics) of the comparison of the denoising algorithms
is provided by calculating the Peak Signal to Noise Ratio (PSNR), the Mean Square
Error (MSE) value and the Mean Absolute Error (MAE) evaluation factors. A
combination of metrics including the PSNR, MSE, and MAE are often required to
clearly assess the model performance
Finding faint HI structure in and around galaxies: scraping the barrel
Soon to be operational HI survey instruments such as APERTIF and ASKAP will
produce large datasets. These surveys will provide information about the HI in
and around hundreds of galaxies with a typical signal-to-noise ratio of
10 in the inner regions and 1 in the outer regions. In addition, such
surveys will make it possible to probe faint HI structures, typically located
in the vicinity of galaxies, such as extra-planar-gas, tails and filaments.
These structures are crucial for understanding galaxy evolution, particularly
when they are studied in relation to the local environment. Our aim is to find
optimized kernels for the discovery of faint and morphologically complex HI
structures. Therefore, using HI data from a variety of galaxies, we explore
state-of-the-art filtering algorithms. We show that the intensity-driven
gradient filter, due to its adaptive characteristics, is the optimal choice. In
fact, this filter requires only minimal tuning of the input parameters to
enhance the signal-to-noise ratio of faint components. In addition, it does not
degrade the resolution of the high signal-to-noise component of a source. The
filtering process must be fast and be embedded in an interactive visualization
tool in order to support fast inspection of a large number of sources. To
achieve such interactive exploration, we implemented a multi-core CPU (OpenMP)
and a GPU (OpenGL) version of this filter in a 3D visualization environment
().Comment: 17 pages, 9 figures, 4 tables. Astronomy and Computing, accepte
- …