4,360 research outputs found

    Bayesian demosaicing using Gaussian scale mixture priors with local adaptivity in the dual tree complex wavelet packet transform domain

    Get PDF
    In digital cameras and mobile phones, there is an ongoing trend to increase the image resolution, decrease the sensor size and to use lower exposure times. Because smaller sensors inherently lead to more noise and a worse spatial resolution, digital post-processing techniques are required to resolve many of the artifacts. Color filter arrays (CFAs), which use alternating patterns of color filters, are very popular because of price and power consumption reasons. However, color filter arrays require the use of a post-processing technique such as demosaicing to recover full resolution RGB images. Recently, there has been some interest in techniques that jointly perform the demosaicing and denoising. This has the advantage that the demosaicing and denoising can be performed optimally (e.g. in the MSE sense) for the considered noise model, while avoiding artifacts introduced when using demosaicing and denoising sequentially. ABSTRACT In this paper, we will continue the research line of the wavelet-based demosaicing techniques. These approaches are computationally simple and very suited for combination with denoising. Therefore, we will derive Bayesian Minimum Squared Error (MMSE) joint demosaicing and denoising rules in the complex wavelet packet domain, taking local adaptivity into account. As an image model, we will use Gaussian Scale Mixtures, thereby taking advantage of the directionality of the complex wavelets. Our results show that this technique is well capable of reconstructing fine details in the image, while removing all of the noise, at a relatively low computational cost. In particular, the complete reconstruction (including color correction, white balancing etc) of a 12 megapixel RAW image takes 3.5 sec on a recent mid-range GPU

    Multi-GPU maximum entropy image synthesis for radio astronomy

    Full text link
    The maximum entropy method (MEM) is a well known deconvolution technique in radio-interferometry. This method solves a non-linear optimization problem with an entropy regularization term. Other heuristics such as CLEAN are faster but highly user dependent. Nevertheless, MEM has the following advantages: it is unsupervised, it has a statistical basis, it has a better resolution and better image quality under certain conditions. This work presents a high performance GPU version of non-gridding MEM, which is tested using real and simulated data. We propose a single-GPU and a multi-GPU implementation for single and multi-spectral data, respectively. We also make use of the Peer-to-Peer and Unified Virtual Addressing features of newer GPUs which allows to exploit transparently and efficiently multiple GPUs. Several ALMA data sets are used to demonstrate the effectiveness in imaging and to evaluate GPU performance. The results show that a speedup from 1000 to 5000 times faster than a sequential version can be achieved, depending on data and image size. This allows to reconstruct the HD142527 CO(6-5) short baseline data set in 2.1 minutes, instead of 2.5 days that takes a sequential version on CPU.Comment: 11 pages, 13 figure

    Analysis, filtering, and control for Takagi-Sugeno fuzzy models in networked systems

    Get PDF
    Copyright © 2015 Sunjie Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.The fuzzy logic theory has been proven to be effective in dealing with various nonlinear systems and has a great success in industry applications. Among different kinds of models for fuzzy systems, the so-called Takagi-Sugeno (T-S) fuzzy model has been quite popular due to its convenient and simple dynamic structure as well as its capability of approximating any smooth nonlinear function to any specified accuracy within any compact set. In terms of such a model, the performance analysis and the design of controllers and filters play important roles in the research of fuzzy systems. In this paper, we aim to survey some recent advances on the T-S fuzzy control and filtering problems with various network-induced phenomena. The network-induced phenomena under consideration mainly include communication delays, packet dropouts, signal quantization, and randomly occurring uncertainties (ROUs). With such network-induced phenomena, the developments on T-S fuzzy control and filtering issues are reviewed in detail. In addition, some latest results on this topic are highlighted. In the end, conclusions are drawn and some possible future research directions are pointed out.This work was supported in part by the National Natural Science Foundation of China under Grants 61134009, 61329301, 11301118 and 61174136, the Natural Science Foundation of Jiangsu Province of China under Grant BK20130017, the Fundamental Research Funds for the Central Universities of China under Grant CUSF-DH-D-2013061, the Royal Society of the U.K., and the Alexander von Humboldt Foundation of Germany

    Adaptive Windowed Sinc Filter for Image Interpolation

    Get PDF
    In this thesis, an image interpolation scheme using an adaptive windowed sinc, which varies in length and frequency response guided by the Laplacian operator, is proposed to improve quality and accuracy. First, an optimization is performed to minimize the squared error for each sliding window by sweeping a range of windowed sinc filters with varying filter lengths. The relationship between the windows, filter length, and edge intensity is analyzed to determine the optimal beta value and filter length. This analysis inspires an image interpolation approach that uses the Laplacian of each sliding window to choose the optimal beta and filter length. The performance of the proposed approach is compared with the optimized result, as well as traditional interpolation methods such as bilinear and bicubic, in terms of PSNR, SSIM, and a subjective visual test. The findings demonstrate the effectiveness of the proposed method in enhancing the resolution of images, contributing to a deeper understanding of adaptive windowed sinc filters and their potential applications in image interpolation

    Two dimensional recursive digital filters for near real time image processing

    Get PDF
    A program was designed toward the demonstration of the feasibility of using two dimensional recursive digital filters for subjective image processing applications that require rapid turn around. The concept of the use of a dedicated minicomputer for the processor for this application was demonstrated. The minicomputer used was the HP1000 series E with a RTE 2 disc operating system and 32K words of memory. A Grinnel 256 x 512 x 8 bit display system was used to display the images. Sample images were provided by NASA Goddard on a 800 BPI, 9 track tape. Four 512 x 512 images representing 4 spectral regions of the same scene were provided. These images were filtered with enhancement filters developed during this effort

    Basic Filters for Convolutional Neural Networks Applied to Music: Training or Design?

    Full text link
    When convolutional neural networks are used to tackle learning problems based on music or, more generally, time series data, raw one-dimensional data are commonly pre-processed to obtain spectrogram or mel-spectrogram coefficients, which are then used as input to the actual neural network. In this contribution, we investigate, both theoretically and experimentally, the influence of this pre-processing step on the network's performance and pose the question, whether replacing it by applying adaptive or learned filters directly to the raw data, can improve learning success. The theoretical results show that approximately reproducing mel-spectrogram coefficients by applying adaptive filters and subsequent time-averaging is in principle possible. We also conducted extensive experimental work on the task of singing voice detection in music. The results of these experiments show that for classification based on Convolutional Neural Networks the features obtained from adaptive filter banks followed by time-averaging perform better than the canonical Fourier-transform-based mel-spectrogram coefficients. Alternative adaptive approaches with center frequencies or time-averaging lengths learned from training data perform equally well.Comment: Completely revised version; 21 pages, 4 figure
    • 

    corecore