3,914 research outputs found
MDL Denoising Revisited
We refine and extend an earlier MDL denoising criterion for wavelet-based
denoising. We start by showing that the denoising problem can be reformulated
as a clustering problem, where the goal is to obtain separate clusters for
informative and non-informative wavelet coefficients, respectively. This
suggests two refinements, adding a code-length for the model index, and
extending the model in order to account for subband-dependent coefficient
distributions. A third refinement is derivation of soft thresholding inspired
by predictive universal coding with weighted mixtures. We propose a practical
method incorporating all three refinements, which is shown to achieve good
performance and robustness in denoising both artificial and natural signals.Comment: Submitted to IEEE Transactions on Information Theory, June 200
Using the Expectation Maximization Algorithm with Heterogeneous Mixture Components for the Analysis of Spectrometry Data
Coupling a multi-capillary column (MCC) with an ion mobility (IM)
spectrometer (IMS) opened a multitude of new application areas for gas
analysis, especially in a medical context, as volatile organic compounds (VOCs)
in exhaled breath can hint at a person's state of health. To obtain a potential
diagnosis from a raw MCC/IMS measurement, several computational steps are
necessary, which so far have required manual interaction, e.g., human
evaluation of discovered peaks. We have recently proposed an automated pipeline
for this task that does not require human intervention during the analysis.
Nevertheless, there is a need for improved methods for each computational step.
In comparison to gas chromatography / mass spectrometry (GC/MS) data, MCC/IMS
data is easier and less expensive to obtain, but peaks are more diffuse and
there is a higher noise level. MCC/IMS measurements can be described as samples
of mixture models (i.e., of convex combinations) of two-dimensional probability
distributions. So we use the expectation-maximization (EM) algorithm to
deconvolute mixtures in order to develop methods that improve data processing
in three computational steps: denoising, baseline correction and peak
clustering. A common theme of these methods is that mixture components within
one model are not homogeneous (e.g., all Gaussian), but of different types.
Evaluation shows that the novel methods outperform the existing ones. We
provide Python software implementing all three methods and make our evaluation
data available at http://www.rahmannlab.de/research/ims
Grounding semantics in robots for Visual Question Answering
In this thesis I describe an operational implementation of an object detection and description system that incorporates in an end-to-end Visual Question Answering system and evaluated it on two visual question answering datasets for compositional language and elementary visual reasoning
Geodesics on the manifold of multivariate generalized Gaussian distributions with an application to multicomponent texture discrimination
We consider the Rao geodesic distance (GD) based on the Fisher information as a similarity measure on the manifold of zero-mean multivariate generalized Gaussian distributions (MGGD). The MGGD is shown to be an adequate model for the heavy-tailed wavelet statistics in multicomponent images, such as color or multispectral images. We discuss the estimation of MGGD parameters using various methods. We apply the GD between MGGDs to color texture discrimination in several classification experiments, taking into account the correlation structure between the spectral bands in the wavelet domain. We compare the performance, both in terms of texture discrimination capability and computational load, of the GD and the Kullback-Leibler divergence (KLD). Likewise, both uni- and multivariate generalized Gaussian models are evaluated, characterized by a fixed or a variable shape parameter. The modeling of the interband correlation significantly improves classification efficiency, while the GD is shown to consistently outperform the KLD as a similarity measure
- …