4,988 research outputs found
On Graphical Models via Univariate Exponential Family Distributions
Undirected graphical models, or Markov networks, are a popular class of
statistical models, used in a wide variety of applications. Popular instances
of this class include Gaussian graphical models and Ising models. In many
settings, however, it might not be clear which subclass of graphical models to
use, particularly for non-Gaussian and non-categorical data. In this paper, we
consider a general sub-class of graphical models where the node-wise
conditional distributions arise from exponential families. This allows us to
derive multivariate graphical model distributions from univariate exponential
family distributions, such as the Poisson, negative binomial, and exponential
distributions. Our key contributions include a class of M-estimators to fit
these graphical model distributions; and rigorous statistical analysis showing
that these M-estimators recover the true graphical model structure exactly,
with high probability. We provide examples of genomic and proteomic networks
learned via instances of our class of graphical models derived from Poisson and
exponential distributions.Comment: Journal of Machine Learning Researc
mgm: Estimating Time-Varying Mixed Graphical Models in High-Dimensional Data
We present the R-package mgm for the estimation of k-order Mixed Graphical
Models (MGMs) and mixed Vector Autoregressive (mVAR) models in high-dimensional
data. These are a useful extensions of graphical models for only one variable
type, since data sets consisting of mixed types of variables (continuous,
count, categorical) are ubiquitous. In addition, we allow to relax the
stationarity assumption of both models by introducing time-varying versions
MGMs and mVAR models based on a kernel weighting approach. Time-varying models
offer a rich description of temporally evolving systems and allow to identify
external influences on the model structure such as the impact of interventions.
We provide the background of all implemented methods and provide fully
reproducible examples that illustrate how to use the package
On a representation of time space-harmonic polynomials via symbolic L\'evy processes
In this paper, we review the theory of time space-harmonic polynomials
developed by using a symbolic device known in the literature as the classical
umbral calculus. The advantage of this symbolic tool is twofold. First a moment
representation is allowed for a wide class of polynomial stochastic involving
the L\'evy processes in respect to which they are martingales. This
representation includes some well-known examples such as Hermite polynomials in
connection with Brownian motion. As a consequence, characterizations of many
other families of polynomials having the time space-harmonic property can be
recovered via the symbolic moment representation. New relations with
Kailath-Segall polynomials are stated. Secondly the generalization to the
multivariable framework is straightforward. Connections with cumulants and Bell
polynomials are highlighted both in the univariate case and in the multivariate
one. Open problems are addressed at the end of the paper
Propriety of Posteriors in Structured Additive Regression Models: Theory and Empirical Evidence
Structured additive regression comprises many semiparametric regression models such as generalized additive (mixed) models, geoadditive models, and hazard regression models within a unified framework. In a Bayesian formulation, nonparametric functions, spatial effects and further model components are specified in terms of multivariate Gaussian priors for high-dimensional vectors of regression coefficients. For several model terms, such as penalised splines or Markov random fields, these Gaussian prior distributions involve rank-deficient precision matrices, yielding partially improper priors. Moreover, hyperpriors for the variances (corresponding to inverse smoothing parameters) may also be specified as improper, e.g. corresponding to Jeffery's prior or a flat prior for the standard deviation. Hence, propriety of the joint posterior is a crucial issue for full Bayesian inference in particular if based on Markov chain Monte Carlo simulations. We establish theoretical results providing sufficient (and sometimes necessary) conditions for propriety and provide empirical evidence through several accompanying simulation studies
Conditional Sum-Product Networks: Imposing Structure on Deep Probabilistic Architectures
Probabilistic graphical models are a central tool in AI; however, they are
generally not as expressive as deep neural models, and inference is notoriously
hard and slow. In contrast, deep probabilistic models such as sum-product
networks (SPNs) capture joint distributions in a tractable fashion, but still
lack the expressive power of intractable models based on deep neural networks.
Therefore, we introduce conditional SPNs (CSPNs), conditional density
estimators for multivariate and potentially hybrid domains which allow
harnessing the expressive power of neural networks while still maintaining
tractability guarantees. One way to implement CSPNs is to use an existing SPN
structure and condition its parameters on the input, e.g., via a deep neural
network. This approach, however, might misrepresent the conditional
independence structure present in data. Consequently, we also develop a
structure-learning approach that derives both the structure and parameters of
CSPNs from data. Our experimental evidence demonstrates that CSPNs are
competitive with other probabilistic models and yield superior performance on
multilabel image classification compared to mean field and mixture density
networks. Furthermore, they can successfully be employed as building blocks for
structured probabilistic models, such as autoregressive image models.Comment: 13 pages, 6 figure
- …