2,202 research outputs found
Sparse Modeling for Image and Vision Processing
In recent years, a large amount of multi-disciplinary research has been
conducted on sparse models and their applications. In statistics and machine
learning, the sparsity principle is used to perform model selection---that is,
automatically selecting a simple model among a large collection of them. In
signal processing, sparse coding consists of representing data with linear
combinations of a few dictionary elements. Subsequently, the corresponding
tools have been widely adopted by several scientific communities such as
neuroscience, bioinformatics, or computer vision. The goal of this monograph is
to offer a self-contained view of sparse modeling for visual recognition and
image processing. More specifically, we focus on applications where the
dictionary is learned and adapted to data, yielding a compact representation
that has been successful in various contexts.Comment: 205 pages, to appear in Foundations and Trends in Computer Graphics
and Visio
Theory and Applications of Robust Optimization
In this paper we survey the primary research, both theoretical and applied,
in the area of Robust Optimization (RO). Our focus is on the computational
attractiveness of RO approaches, as well as the modeling power and broad
applicability of the methodology. In addition to surveying prominent
theoretical results of RO, we also present some recent results linking RO to
adaptable models for multi-stage decision-making problems. Finally, we
highlight applications of RO across a wide spectrum of domains, including
finance, statistics, learning, and various areas of engineering.Comment: 50 page
Nonconcave penalized likelihood with a diverging number of parameters
A class of variable selection procedures for parametric models via nonconcave
penalized likelihood was proposed by Fan and Li to simultaneously estimate
parameters and select important variables. They demonstrated that this class of
procedures has an oracle property when the number of parameters is finite.
However, in most model selection problems the number of parameters should be
large and grow with the sample size. In this paper some asymptotic properties
of the nonconcave penalized likelihood are established for situations in which
the number of parameters tends to \infty as the sample size increases.
Under regularity conditions we have established an oracle property and the
asymptotic normality of the penalized likelihood estimators. Furthermore, the
consistency of the sandwich formula of the covariance matrix is demonstrated.
Nonconcave penalized likelihood ratio statistics are discussed, and their
asymptotic distributions under the null hypothesis are obtained by imposing
some mild conditions on the penalty functions
Bayesian inference for inverse problems
Traditionally, the MaxEnt workshops start by a tutorial day. This paper
summarizes my talk during 2001'th workshop at John Hopkins University. The main
idea in this talk is to show how the Bayesian inference can naturally give us
all the necessary tools we need to solve real inverse problems: starting by
simple inversion where we assume to know exactly the forward model and all the
input model parameters up to more realistic advanced problems of myopic or
blind inversion where we may be uncertain about the forward model and we may
have noisy data. Starting by an introduction to inverse problems through a few
examples and explaining their ill posedness nature, I briefly presented the
main classical deterministic methods such as data matching and classical
regularization methods to show their limitations. I then presented the main
classical probabilistic methods based on likelihood, information theory and
maximum entropy and the Bayesian inference framework for such problems. I show
that the Bayesian framework, not only generalizes all these methods, but also
gives us natural tools, for example, for inferring the uncertainty of the
computed solutions, for the estimation of the hyperparameters or for handling
myopic or blind inversion problems. Finally, through a deconvolution problem
example, I presented a few state of the art methods based on Bayesian inference
particularly designed for some of the mass spectrometry data processing
problems.Comment: Presented at MaxEnt01. To appear in Bayesian Inference and Maximum
Entropy Methods, B. Fry (Ed.), AIP Proceedings. 20pages, 13 Postscript
figure
- …