8,135 research outputs found
Inference via low-dimensional couplings
We investigate the low-dimensional structure of deterministic transformations
between random variables, i.e., transport maps between probability measures. In
the context of statistics and machine learning, these transformations can be
used to couple a tractable "reference" measure (e.g., a standard Gaussian) with
a target measure of interest. Direct simulation from the desired measure can
then be achieved by pushing forward reference samples through the map. Yet
characterizing such a map---e.g., representing and evaluating it---grows
challenging in high dimensions. The central contribution of this paper is to
establish a link between the Markov properties of the target measure and the
existence of low-dimensional couplings, induced by transport maps that are
sparse and/or decomposable. Our analysis not only facilitates the construction
of transformations in high-dimensional settings, but also suggests new
inference methodologies for continuous non-Gaussian graphical models. For
instance, in the context of nonlinear state-space models, we describe new
variational algorithms for filtering, smoothing, and sequential parameter
inference. These algorithms can be understood as the natural
generalization---to the non-Gaussian case---of the square-root
Rauch-Tung-Striebel Gaussian smoother.Comment: 78 pages, 25 figure
Efficient Learning of Sparse Conditional Random Fields for Supervised Sequence Labelling
Conditional Random Fields (CRFs) constitute a popular and efficient approach
for supervised sequence labelling. CRFs can cope with large description spaces
and can integrate some form of structural dependency between labels. In this
contribution, we address the issue of efficient feature selection for CRFs
based on imposing sparsity through an L1 penalty. We first show how sparsity of
the parameter set can be exploited to significantly speed up training and
labelling. We then introduce coordinate descent parameter update schemes for
CRFs with L1 regularization. We finally provide some empirical comparisons of
the proposed approach with state-of-the-art CRF training strategies. In
particular, it is shown that the proposed approach is able to take profit of
the sparsity to speed up processing and hence potentially handle larger
dimensional models
A latent discriminative model-based approach for classification of imaginary motor tasks from EEG data
We consider the problem of classification of imaginary motor tasks from
electroencephalography (EEG) data for brain-computer interfaces (BCIs) and propose a new approach based on hidden conditional random fields (HCRFs). HCRFs are discriminative graphical models that are attractive for this problem because they (1) exploit the temporal structure of EEG; (2) include latent variables that can be used to model different brain states in the signal; and (3) involve learned statistical models matched to the classification task, avoiding some of the limitations of generative models. Our approach involves spatial filtering of the EEG signals and estimation of power spectra based on auto-regressive modeling of temporal segments of the EEG signals. Given this time-frequency representation, we select certain frequency bands that are known to be associated with execution of motor tasks. These selected features constitute the data that are fed to the HCRF, parameters of which are learned from training data. Inference algorithms on the HCRFs are used for classification of motor tasks. We experimentally compare this approach to the best performing methods in BCI competition IV as well as a number of more recent methods and observe that our proposed method yields better classification accuracy
Latent Gaussian modeling and INLA: A review with focus on space-time applications
Bayesian hierarchical models with latent Gaussian layers have proven very
flexible in capturing complex stochastic behavior and hierarchical structures
in high-dimensional spatial and spatio-temporal data. Whereas simulation-based
Bayesian inference through Markov Chain Monte Carlo may be hampered by slow
convergence and numerical instabilities, the inferential framework of
Integrated Nested Laplace Approximation (INLA) is capable to provide accurate
and relatively fast analytical approximations to posterior quantities of
interest. It heavily relies on the use of Gauss-Markov dependence structures to
avoid the numerical bottleneck of high-dimensional nonsparse matrix
computations. With a view towards space-time applications, we here review the
principal theoretical concepts, model classes and inference tools within the
INLA framework. Important elements to construct space-time models are certain
spatial Mat\'ern-like Gauss-Markov random fields, obtained as approximate
solutions to a stochastic partial differential equation. Efficient
implementation of statistical inference tools for a large variety of models is
available through the INLA package of the R software. To showcase the practical
use of R-INLA and to illustrate its principal commands and syntax, a
comprehensive simulation experiment is presented using simulated non Gaussian
space-time count data with a first-order autoregressive dependence structure in
time
Information-geometric Markov Chain Monte Carlo methods using Diffusions
Recent work incorporating geometric ideas in Markov chain Monte Carlo is
reviewed in order to highlight these advances and their possible application in
a range of domains beyond Statistics. A full exposition of Markov chains and
their use in Monte Carlo simulation for Statistical inference and molecular
dynamics is provided, with particular emphasis on methods based on Langevin
diffusions. After this geometric concepts in Markov chain Monte Carlo are
introduced. A full derivation of the Langevin diffusion on a Riemannian
manifold is given, together with a discussion of appropriate Riemannian metric
choice for different problems. A survey of applications is provided, and some
open questions are discussed.Comment: 22 pages, 2 figure
- …