30,058 research outputs found

    A General Method for Amortizing Variational Filtering

    Full text link
    We introduce the variational filtering EM algorithm, a simple, general-purpose method for performing variational inference in dynamical latent variable models using information from only past and present variables, i.e. filtering. The algorithm is derived from the variational objective in the filtering setting and consists of an optimization procedure at each time step. By performing each inference optimization procedure with an iterative amortized inference model, we obtain a computationally efficient implementation of the algorithm, which we call amortized variational filtering. We present experiments demonstrating that this general-purpose method improves performance across several deep dynamical latent variable models.Comment: Advances in Neural Information Processing Systems (NIPS) 201

    Music generation with variational recurrent autoencoder supported by history

    Full text link
    A new architecture of an artificial neural network that helps to generate longer melodic patterns is introduced alongside with methods for post-generation filtering. The proposed approach called variational autoencoder supported by history is based on a recurrent highway gated network combined with a variational autoencoder. Combination of this architecture with filtering heuristics allows generating pseudo-live acoustically pleasing and melodically diverse music

    Gaussian filtering and variational approximations for Bayesian smoothing in continuous-discrete stochastic dynamic systems

    Full text link
    The Bayesian smoothing equations are generally intractable for systems described by nonlinear stochastic differential equations and discrete-time measurements. Gaussian approximations are a computationally efficient way to approximate the true smoothing distribution. In this work, we present a comparison between two Gaussian approximation methods. The Gaussian filtering based Gaussian smoother uses a Gaussian approximation for the filtering distribution to form an approximation for the smoothing distribution. The variational Gaussian smoother is based on minimizing the Kullback-Leibler divergence of the approximate smoothing distribution with respect to the true distribution. The results suggest that for highly nonlinear systems, the variational Gaussian smoother can be used to iteratively improve the Gaussian filtering based smoothing solution. We also present linearization and sigma-point methods to approximate the intractable Gaussian expectations in the Variational Gaussian smoothing equations. In addition, we extend the variational Gaussian smoother for certain class of systems with singular diffusion matrix.Comment: 24 pages, 4 figures, submitted to Signal Processin

    Variational Autoencoders for Collaborative Filtering

    Full text link
    We extend variational autoencoders (VAEs) to collaborative filtering for implicit feedback. This non-linear probabilistic model enables us to go beyond the limited modeling capacity of linear factor models which still largely dominate collaborative filtering research.We introduce a generative model with multinomial likelihood and use Bayesian inference for parameter estimation. Despite widespread use in language modeling and economics, the multinomial likelihood receives less attention in the recommender systems literature. We introduce a different regularization parameter for the learning objective, which proves to be crucial for achieving competitive performance. Remarkably, there is an efficient way to tune the parameter using annealing. The resulting model and learning algorithm has information-theoretic connections to maximum entropy discrimination and the information bottleneck principle. Empirically, we show that the proposed approach significantly outperforms several state-of-the-art baselines, including two recently-proposed neural network approaches, on several real-world datasets. We also provide extended experiments comparing the multinomial likelihood with other commonly used likelihood functions in the latent factor collaborative filtering literature and show favorable results. Finally, we identify the pros and cons of employing a principled Bayesian inference approach and characterize settings where it provides the most significant improvements.Comment: 10 pages, 3 figures. WWW 201

    Neural Variational Hybrid Collaborative Filtering

    Full text link
    Collaborative Filtering (CF) is one of the most used methods for Recommender System. Because of the Bayesian nature and nonlinearity, deep generative models, e.g. Variational Autoencoder (VAE), have been applied into CF task, and have achieved great performance. However, most VAE-based methods suffer from matrix sparsity and consider the prior of users' latent factors to be the same, which leads to poor latent representations of users and items. Additionally, most existing methods model latent factors of users only and but not items, which makes them not be able to recommend items to a new user. To tackle these problems, we propose a Neural Variational Hybrid Collaborative Filtering, NVHCF. Specifically, we consider both the generative processes of users and items, and the prior of latent factors of users and items to be side informationspecific, which enables our model to alleviate matrix sparsity and learn better latent representations of users and items. For inference purpose, we derived a Stochastic Gradient Variational Bayes (SGVB) algorithm to analytically approximate the intractable distributions of latent factors of users and items. Experiments conducted on two large datasets have showed our methods significantly outperform the state-of-the-art CF methods, including the VAE-based methods.Comment: 7 pages, fix some typos (Eq.7

    A Variational Formulation of Optimal Nonlinear Estimation

    Get PDF
    We propose a variational method to solve all three estimation problems for nonlinear stochastic dynamical systems: prediction, filtering, and smoothing. Our new approach is based upon a proper choice of cost function, termed the {\it effective action}. We show that this functional of time-histories is the unique statistically well-founded cost function to determine most probable histories within empirical ensembles. The ensemble dispersion about the sample mean history can also be obtained from the Hessian of the cost function. We show that the effective action can be calculated by a variational prescription, which generalizes the ``sweep method'' used in optimal linear estimation. An iterative numerical scheme results which converges globally to the variational estimator. This scheme involves integrating forward in time a ``perturbed'' Fokker-Planck equation, very closely related to the Kushner-Stratonovich equation for optimal filtering, and an adjoint equation backward in time, similarly related to the Pardoux-Kushner equation for optimal smoothing. The variational estimator enjoys a somewhat weaker property, which we call ``mean optimality''. However, the variational scheme has the principal advantage---crucial for practical applications---that it admits a wide variety of finite-dimensional moment-closure approximations. The moment approximations are derived reductively from the Euler-Lagrange variational formulation and preserve the good structural properties of the optimal estimator.Comment: 59 pages. Glaring typo on p.2 corrected, refs update

    Variational Bayesian Adaptation of Noise Covariances in Non-Linear Kalman Filtering

    Full text link
    This paper is considered with joint estimation of state and time-varying noise covariance matrices in non-linear stochastic state space models. We present a variational Bayes and Gaussian filtering based algorithm for efficient computation of the approximate filtering posterior distributions. The Gaussian filtering based formulation of the non-linear state space model computation allows usage of efficient Gaussian integration methods such as unscented transform, cubature integration and Gauss-Hermite integration along with the classical Taylor series approximations. The performance of the algorithm is illustrated in a simulated application

    Switching Linear Dynamics for Variational Bayes Filtering

    Full text link
    System identification of complex and nonlinear systems is a central problem for model predictive control and model-based reinforcement learning. Despite their complexity, such systems can often be approximated well by a set of linear dynamical systems if broken into appropriate subsequences. This mechanism not only helps us find good approximations of dynamics, but also gives us deeper insight into the underlying system. Leveraging Bayesian inference, Variational Autoencoders and Concrete relaxations, we show how to learn a richer and more meaningful state space, e.g. encoding joint constraints and collisions with walls in a maze, from partial and high-dimensional observations. This representation translates into a gain of accuracy of learned dynamics showcased on various simulated tasks.Comment: Appears in Proceedings of the 36th International Conference on Machine Learning (ICML

    Structured Black Box Variational Inference for Latent Time Series Models

    Full text link
    Continuous latent time series models are prevalent in Bayesian modeling; examples include the Kalman filter, dynamic collaborative filtering, or dynamic topic models. These models often benefit from structured, non mean field variational approximations that capture correlations between time steps. Black box variational inference with reparameterization gradients (BBVI) allows us to explore a rich new class of Bayesian non-conjugate latent time series models; however, a naive application of BBVI to such structured variational models would scale quadratically in the number of time steps. We describe a BBVI algorithm analogous to the forward-backward algorithm which instead scales linearly in time. It allows us to efficiently sample from the variational distribution and estimate the gradients of the ELBO. Finally, we show results on the recently proposed dynamic word embedding model, which was trained using our method.Comment: 5 pages, 1 figure; presented at the ICML 2017 Time Series Worksho

    Adaptive Variational Particle Filtering in Non-stationary Environments

    Full text link
    Online convex optimization is a sequential prediction framework with the goal to track and adapt to the environment through evaluating proper convex loss functions. We study efficient particle filtering methods from the perspective of such a framework. We formulate an efficient particle filtering methods for the non-stationary environment by making connections with the online mirror descent algorithm which is known to be a universal online convex optimization algorithm. As a result of this connection, our proposed particle filtering algorithm proves to achieve optimal particle efficiency
    corecore