36,350 research outputs found
Learning probability distributions generated by finite-state machines
We review methods for inference of probability distributions generated by probabilistic automata and related models for sequence generation. We focus on methods that can be proved to learn in the inference
in the limit and PAC formal models. The methods we review are state merging and state splitting methods for probabilistic deterministic automata and the recently developed spectral method for nondeterministic probabilistic automata. In both cases, we derive them from a high-level algorithm described in terms of the Hankel matrix of the distribution to be learned, given as an oracle, and then describe how to adapt that algorithm to account for the error introduced by a finite sample.Peer ReviewedPostprint (author's final draft
Probabilistic error estimation for non-intrusive reduced models learned from data of systems governed by linear parabolic partial differential equations
This work derives a residual-based a posteriori error estimator for reduced
models learned with non-intrusive model reduction from data of high-dimensional
systems governed by linear parabolic partial differential equations with
control inputs. It is shown that quantities that are necessary for the error
estimator can be either obtained exactly as the solutions of least-squares
problems in a non-intrusive way from data such as initial conditions, control
inputs, and high-dimensional solution trajectories or bounded in a
probabilistic sense. The computational procedure follows an offline/online
decomposition. In the offline (training) phase, the high-dimensional system is
judiciously solved in a black-box fashion to generate data and to set up the
error estimator. In the online phase, the estimator is used to bound the error
of the reduced-model predictions for new initial conditions and new control
inputs without recourse to the high-dimensional system. Numerical results
demonstrate the workflow of the proposed approach from data to reduced models
to certified predictions
Symbolic Exact Inference for Discrete Probabilistic Programs
The computational burden of probabilistic inference remains a hurdle for
applying probabilistic programming languages to practical problems of interest.
In this work, we provide a semantic and algorithmic foundation for efficient
exact inference on discrete-valued finite-domain imperative probabilistic
programs. We leverage and generalize efficient inference procedures for
Bayesian networks, which exploit the structure of the network to decompose the
inference task, thereby avoiding full path enumeration. To do this, we first
compile probabilistic programs to a symbolic representation. Then we adapt
techniques from the probabilistic logic programming and artificial intelligence
communities in order to perform inference on the symbolic representation. We
formalize our approach, prove it sound, and experimentally validate it against
existing exact and approximate inference techniques. We show that our inference
approach is competitive with inference procedures specialized for Bayesian
networks, thereby expanding the class of probabilistic programs that can be
practically analyzed
Learning Tree Distributions by Hidden Markov Models
Hidden tree Markov models allow learning distributions for tree structured
data while being interpretable as nondeterministic automata. We provide a
concise summary of the main approaches in literature, focusing in particular on
the causality assumptions introduced by the choice of a specific tree visit
direction. We will then sketch a novel non-parametric generalization of the
bottom-up hidden tree Markov model with its interpretation as a
nondeterministic tree automaton with infinite states.Comment: Accepted in LearnAut2018 worksho
- …