2,832,341 research outputs found
Deep Gaussian Mixture Models
Deep learning is a hierarchical inference method formed by subsequent
multiple layers of learning able to more efficiently describe complex
relationships. In this work, Deep Gaussian Mixture Models are introduced and
discussed. A Deep Gaussian Mixture model (DGMM) is a network of multiple layers
of latent variables, where, at each layer, the variables follow a mixture of
Gaussian distributions. Thus, the deep mixture model consists of a set of
nested mixtures of linear models, which globally provide a nonlinear model able
to describe the data in a very flexible way. In order to avoid
overparameterized solutions, dimension reduction by factor models can be
applied at each layer of the architecture thus resulting in deep mixtures of
factor analysers.Comment: 19 pages, 4 figure
Jeffreys priors for mixture estimation
While Jeffreys priors usually are well-defined for the parameters of mixtures
of distributions, they are not available in closed form. Furthermore, they
often are improper priors. Hence, they have never been used to draw inference
on the mixture parameters. We study in this paper the implementation and the
properties of Jeffreys priors in several mixture settings, show that the
associated posterior distributions most often are improper, and then propose a
noninformative alternative for the analysis of mixtures
- …
