237 research outputs found
Recommended from our members
Time-Frequency Analysis as Probabilistic Inference
This is the final published version. It was originally published by IEEE at http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=6918491.This paper proposes a new view of time-frequency analysis framed in terms of probabilistic inference. Natural signals are assumed to be formed by the superposition of distinct time-frequency components, with the analytic goal being to infer these components by application of Bayes' rule. The framework serves to unify various existing models for natural time-series; it relates to both the Wiener and Kalman filters, and with suitable assumptions yields inferential interpretations of the short-time Fourier transform, spectrogram, filter bank, and wavelet representations. Value is gained by placing time-frequency analysis on the same probabilistic basis as is often employed in applications such as denoising, source separation, or recognition. Uncertainty in the time-frequency representation can be propagated correctly to application-specific stages, improving the handing of noise and missing data. Probabilistic learning allows modules to be co-adapted; thus, the time-frequency representation can be adapted to both the demands of the application and the time-varying statistics of the signal at hand. Similarly, the application module can be adapted to fine properties of the signal propagated by the initial time-frequency processing. We demonstrate these benefits by combining probabilistic time-frequency representations with non-negative matrix factorization, finding benefits in audio denoising and inpainting tasks, albeit with higher computational cost than incurred by the standard approach.Funding was provided by EPSRC (grant numbers EP/G050821/1 and
EP/L000776/1) and Google (R.E.T.) and by the Gatsby Charitable Foundation
(M.S.)
Directional edge and texture representations for image processing
An efficient representation for natural images is of fundamental importance in image processing and analysis. The commonly used separable transforms such as wavelets axe not best suited for images due to their inability to exploit directional regularities such as edges and oriented textural patterns; while most of the recently proposed directional schemes cannot represent these two types of features in a unified transform. This thesis focuses on the development of directional representations for images which can capture both edges and textures in a multiresolution manner. The thesis first considers the problem of extracting linear features with the multiresolution Fourier transform (MFT). Based on a previous MFT-based linear feature model, the work extends the extraction method into the situation when the image is corrupted by noise. The problem is tackled by the combination of a "Signal+Noise" frequency model, a refinement stage and a robust classification scheme. As a result, the MFT is able to perform linear feature analysis on noisy images on which previous methods failed. A new set of transforms called the multiscale polar cosine transforms (MPCT) are also proposed in order to represent textures. The MPCT can be regarded as real-valued MFT with similar basis functions of oriented sinusoids. It is shown that the transform can represent textural patches more efficiently than the conventional Fourier basis. With a directional best cosine basis, the MPCT packet (MPCPT) is shown to be an efficient representation for edges and textures, despite its high computational burden. The problem of representing edges and textures in a fixed transform with less complexity is then considered. This is achieved by applying a Gaussian frequency filter, which matches the disperson of the magnitude spectrum, on the local MFT coefficients. This is particularly effective in denoising natural images, due to its ability to preserve both types of feature. Further improvements can be made by employing the information given by the linear feature extraction process in the filter's configuration. The denoising results compare favourably against other state-of-the-art directional representations
Multiframe visual-inertial blur estimation and removal for unmodified smartphones
Pictures and videos taken with smartphone cameras often suffer from motion blur due to handshake during the
exposure time. Recovering a sharp frame from a blurry one is an ill-posed problem but in smartphone applications
additional cues can aid the solution. We propose a blur removal algorithm that exploits information from subsequent
camera frames and the built-in inertial sensors of an unmodified smartphone. We extend the fast non-blind
uniform blur removal algorithm of Krishnan and Fergus to non-uniform blur and to multiple input frames. We estimate
piecewise uniform blur kernels from the gyroscope measurements of the smartphone and we adaptively steer
our multiframe deconvolution framework towards the sharpest input patches. We show in qualitative experiments
that our algorithm can remove synthetic and real blur from individual frames of a degraded image sequence within
a few seconds
Computational Imaging Approach to Recovery of Target Coordinates Using Orbital Sensor Data
This dissertation addresses the components necessary for simulation of an image-based recovery of the position of a target using orbital image sensors. Each component is considered in detail, focusing on the effect that design choices and system parameters have on the accuracy of the position estimate. Changes in sensor resolution, varying amounts of blur, differences in image noise level, selection of algorithms used for each component, and lag introduced by excessive processing time all contribute to the accuracy of the result regarding recovery of target coordinates using orbital sensor data.
Using physical targets and sensors in this scenario would be cost-prohibitive in the exploratory setting posed, therefore a simulated target path is generated using Bezier curves which approximate representative paths followed by the targets of interest. Orbital trajectories for the sensors are designed on an elliptical model representative of the motion of physical orbital sensors. Images from each sensor are simulated based on the position and orientation of the sensor, the position of the target, and the imaging parameters selected for the experiment (resolution, noise level, blur level, etc.). Post-processing of the simulated imagery seeks to reduce noise and blur and increase resolution. The only information available for calculating the target position by a fully implemented system are the sensor position and orientation vectors and the images from each sensor. From these data we develop a reliable method of recovering the target position and analyze the impact on near-realtime processing. We also discuss the influence of adjustments to system components on overall capabilities and address the potential system size, weight, and power requirements from realistic implementation approaches
Techniques for enhancing digital images
The images obtain from either research studies or optical instruments are
often corrupted with noise. Image denoising involves the manipulation of image
data to produce a visually high quality image. This thesis reviews the existing
denoising algorithms and the filtering approaches available for enhancing images
and/or data transmission.
Spatial-domain and Transform-domain digital image filtering algorithms
have been used in the past to suppress different noise models. The different noise
models can be either additive or multiplicative. Selection of the denoising algorithm
is application dependent. It is necessary to have knowledge about the noise present
in the image so as to select the appropriated denoising algorithm. Noise models
may include Gaussian noise, Salt and Pepper noise, Speckle noise and Brownian
noise. The Wavelet Transform is similar to the Fourier transform with a completely
different merit function. The main difference between Wavelet transform and
Fourier transform is that, in the Wavelet Transform, Wavelets are localized in both
time and frequency. In the standard Fourier Transform, Wavelets are only localized
in frequency. Wavelet analysis consists of breaking up the signal into shifted and
scales versions of the original (or mother) Wavelet. The Wiener Filter (mean
squared estimation error) finds implementations as a LMS filter (least mean
squares), RLS filter (recursive least squares), or Kalman filter.
Quantitative measure (metrics) of the comparison of the denoising algorithms
is provided by calculating the Peak Signal to Noise Ratio (PSNR), the Mean Square
Error (MSE) value and the Mean Absolute Error (MAE) evaluation factors. A
combination of metrics including the PSNR, MSE, and MAE are often required to
clearly assess the model performance
- …