761 research outputs found
Recommended from our members
Efficient Variational Inference for Hierarchical Models of Images, Text, and Networks
Variational inference provides a general optimization framework to approximate the posterior distributions of latent variables in probabilistic models. Although effective in simple scenarios, variational inference may be inaccurate or infeasible when the data is high-dimensional, the model structure is complicated, or variable relationships are non-conjugate. We propose solutions to these problems through the smart design and leverage of model structures, the rigorous derivation of variational bounds, and the creation of flexible algorithms for various models with rich, non-conjugate dependencies.Concretely, we first design an interpretable generative model for natural images, in which the hundreds of thousands of pixels per image are split into small patches represented by Gaussian mixture models. Through structured variational inference, the evidence lower bound of this model automatically recovers the popular expected patch log-likelihood method for image processing. A nonparametric extension using hierarchical Dirichlet processes further enables self-similarities to be captured and image-specific clusters created during inference, boosting image denoising and inpainting accuracy.Then we move on to text data, and design hierarchical topic graphs that generalize the bipartite noisy-OR models previously used for medical diagnosis. We derive auxiliary bounds to overcome the non-conjugacy of noisy-OR conditionals, and use stochastic variational inference to efficiently train on datasets with hundreds of thousands of documents. We dramatically increase the algorithm speed through a constrained family of variational bounds, so that only the ancestors of the sparse observed tokens of each document need to be considered.Finally, we propose a general-purpose Monte Carlo variational inference strategy that is directly applicable to any model with discrete variables. Compared to REINFORCE-style stochastic gradient updates, our coordinate-ascent updates have lower variance and converge much faster. Compared to auxiliary-variable bounds crafted for each individual model, our algorithm is simpler to derive and may be easily integrated into probabilistic programming languages for broader use. By avoiding auxiliary variables, we also tighten likelihood bounds and increase robustness to local optima. Extensive experiments on real-world models of images, text, and networks illustrate these appealing advantages
Advances in Optimization and Nonlinear Analysis
The present book focuses on that part of calculus of variations, optimization, nonlinear analysis and related applications which combines tools and methods from partial differential equations with geometrical techniques. More precisely, this work is devoted to nonlinear problems coming from different areas, with particular reference to those introducing new techniques capable of solving a wide range of problems. The book is a valuable guide for researchers, engineers and students in the field of mathematics, operations research, optimal control science, artificial intelligence, management science and economics
International Conference on Continuous Optimization (ICCOPT) 2019 Conference Book
The Sixth International Conference on Continuous Optimization took place on the campus of the Technical University of Berlin, August 3-8, 2019. The ICCOPT is a flagship conference of the Mathematical Optimization Society (MOS), organized every three years. ICCOPT 2019 was hosted by the Weierstrass Institute for Applied Analysis and Stochastics (WIAS) Berlin. It included a Summer School and a Conference with a series of plenary and semi-plenary talks, organized and contributed sessions, and poster sessions.
This book comprises the full conference program. It contains, in particular, the scientific program in survey style as well as with all details, and information on the social program, the venue, special meetings, and more
A primer on noise-induced transitions in applied dynamical systems
Noise plays a fundamental role in a wide variety of physical and biological
dynamical systems. It can arise from an external forcing or due to random
dynamics internal to the system. It is well established that even weak noise
can result in large behavioral changes such as transitions between or escapes
from quasi-stable states. These transitions can correspond to critical events
such as failures or extinctions that make them essential phenomena to
understand and quantify, despite the fact that their occurrence is rare. This
article will provide an overview of the theory underlying the dynamics of rare
events for stochastic models along with some example applications
Curvature-Independent Last-Iterate Convergence for Games on Riemannian Manifolds
Numerous applications in machine learning and data analytics can be
formulated as equilibrium computation over Riemannian manifolds. Despite the
extensive investigation of their Euclidean counterparts, the performance of
Riemannian gradient-based algorithms remain opaque and poorly understood. We
revisit the original scheme of Riemannian gradient descent (RGD) and analyze it
under a geodesic monotonicity assumption, which includes the well-studied
geodesically convex-concave min-max optimization problem as a special case. Our
main contribution is to show that, despite the phenomenon of distance
distortion, the RGD scheme, with a step size that is agnostic to the manifold's
curvature, achieves a curvature-independent and linear last-iterate convergence
rate in the geodesically strongly monotone setting. To the best of our
knowledge, the possibility of curvature-independent rates and/or last-iterate
convergence in the Riemannian setting has not been considered before
- …