143,836 research outputs found
A probabilistic approach to case-based inference
The central problem in case based reasoning (CBR) is to infer a solution for a new problem-instance by using a collection of existing problem-solution cases. The basic heuristic guiding CBR is the hypothesis that similar problems have similar solutions. Recently, some attempts at formalizing CBR in a theoretical framework have been made, including work by Hullermeier who established a link between CBR and the probably approximately correct (PAC) theoretical model of learning in his 'case-based inference' (CBI) formulation. In this paper we develop further such probabilistic modelling, framing CBI it as a multi-category classification problem. We use a recently-developed notion of geometric margin of classification to obtain generalization error bounds
Inferring dynamic genetic networks with low order independencies
In this paper, we propose a novel inference method for dynamic genetic
networks which makes it possible to face with a number of time measurements n
much smaller than the number of genes p. The approach is based on the concept
of low order conditional dependence graph that we extend here in the case of
Dynamic Bayesian Networks. Most of our results are based on the theory of
graphical models associated with the Directed Acyclic Graphs (DAGs). In this
way, we define a minimal DAG G which describes exactly the full order
conditional dependencies given the past of the process. Then, to face with the
large p and small n estimation case, we propose to approximate DAG G by
considering low order conditional independencies. We introduce partial qth
order conditional dependence DAGs G(q) and analyze their probabilistic
properties. In general, DAGs G(q) differ from DAG G but still reflect relevant
dependence facts for sparse networks such as genetic networks. By using this
approximation, we set out a non-bayesian inference method and demonstrate the
effectiveness of this approach on both simulated and real data analysis. The
inference procedure is implemented in the R package 'G1DBN' freely available
from the CRAN archive
Dynamic Inference in Probabilistic Graphical Models
Probabilistic graphical models, such as Markov random fields (MRFs), are
useful for describing high-dimensional distributions in terms of local
dependence structures. The probabilistic inference is a fundamental problem
related to graphical models, and sampling is a main approach for the problem.
In this paper, we study probabilistic inference problems when the graphical
model itself is changing dynamically with time. Such dynamic inference problems
arise naturally in today's application, e.g.~multivariate time-series data
analysis and practical learning procedures.
We give a dynamic algorithm for sampling-based probabilistic inferences in
MRFs, where each dynamic update can change the underlying graph and all
parameters of the MRF simultaneously, as long as the total amount of changes is
bounded. More precisely, suppose that the MRF has variables and
polylogarithmic-bounded maximum degree, and independent samples are
sufficient for the inference for a polynomial function . Our
algorithm dynamically maintains an answer to the inference problem using
space cost, and incremental
time cost upon each update to the MRF, as long as the well-known
Dobrushin-Shlosman condition is satisfied by the MRFs. Compared to the static
case, which requires time cost for redrawing all
samples whenever the MRF changes, our dynamic algorithm gives a
-factor speedup. Our approach relies on a
novel dynamic sampling technique, which transforms local Markov chains (a.k.a.
single-site dynamics) to dynamic sampling algorithms, and an "algorithmic
Lipschitz" condition that we establish for sampling from graphical models,
namely, when the MRF changes by a small difference, samples can be modified to
reflect the new distribution, with cost proportional to the difference on MRF
Approximate inference methods in probabilistic machine learning and Bayesian statistics
This thesis develops new methods for efficient approximate inference in probabilistic models. Such models are routinely used in different fields, yet they remain computationally challenging as they involve high-dimensional integrals. We propose different approximate inference approaches addressing some challenges in probabilistic machine learning and Bayesian statistics. First, we present a Bayesian framework for genome-wide inference of DNA methylation levels and devise an efficient particle filtering and smoothing algorithm that can be used to identify differentially methylated regions between case and control groups. Second, we present a scalable inference approach for state space models by combining variational methods with sequential Monte Carlo sampling. The method is applied to self-exciting point process models that allow for flexible dynamics in the latent intensity function. Third, a new variational density motivated by copulas is developed. This new variational family can be beneficial compared with Gaussian approximations, as illustrated on examples with Bayesian neural networks. Lastly, we make some progress in a gradient-based adaptation of Hamiltonian Monte Carlo samplers by maximizing an approximation of the proposal entropy
- …