17 research outputs found
RankPL: A Qualitative Probabilistic Programming Language
In this paper we introduce RankPL, a modeling language that can be thought of
as a qualitative variant of a probabilistic programming language with a
semantics based on Spohn's ranking theory. Broadly speaking, RankPL can be used
to represent and reason about processes that exhibit uncertainty expressible by
distinguishing "normal" from" surprising" events. RankPL allows (iterated)
revision of rankings over alternative program states and supports various types
of reasoning, including abduction and causal inference. We present the
language, its denotational semantics, and a number of practical examples. We
also discuss an implementation of RankPL that is available for download
Combining Stochastic Constraint Optimization and Probabilistic Programming
Algorithms and the Foundations of Software technolog
Tuning structure learning algorithms with out-of-sample and resampling strategies
One of the challenges practitioners face when applying structure learning algorithms to their data involves determining a set of hyperparameters; otherwise, a set of hyperparameter defaults is assumed. The optimal hyperparameter configuration often depends on multiple factors, including the size and density of the usually unknown underlying true graph, the sample size of the input data, and the structure learning algorithm. We propose a novel hyperparameter tuning method, called the Out-of-sample Tuning for Structure Learning (OTSL), that employs out-of-sample and resampling strategies to estimate the optimal hyperparameter configuration for structure learning, given the input dataset and structure learning algorithm. Synthetic experiments show that employing OTSL to tune the hyperparameters of hybrid and score-based structure learning algorithms leads to improvements in graphical accuracy compared to the state-of-the-art. We also illustrate the applicability of this approach to real datasets from different disciplines
A Proximal Approach for a Class of Matrix Optimization Problems
In recent years, there has been a growing interest in mathematical models
leading to the minimization, in a symmetric matrix space, of a Bregman
divergence coupled with a regularization term. We address problems of this type
within a general framework where the regularization term is split in two parts,
one being a spectral function while the other is arbitrary. A Douglas-Rachford
approach is proposed to address such problems and a list of proximity operators
is provided allowing us to consider various choices for the fit-to-data
functional and for the regularization term. Numerical experiments show the
validity of this approach for solving convex optimization problems encountered
in the context of sparse covariance matrix estimation. Based on our theoretical
results, an algorithm is also proposed for noisy graphical lasso where a
precision matrix has to be estimated in the presence of noise. The nonconvexity
of the resulting objective function is dealt with a majorization-minimization
approach, i.e. by building a sequence of convex surrogates and solving the
inner optimization subproblems via the aforementioned Douglas-Rachford
procedure. We establish conditions for the convergence of this iterative scheme
and we illustrate its good numerical performance with respect to
state-of-the-art approaches
Exact Learning Augmented Naive Bayes Classifier
Earlier studies have shown that classification accuracies of Bayesian
networks (BNs) obtained by maximizing the conditional log likelihood (CLL) of a
class variable, given the feature variables, were higher than those obtained by
maximizing the marginal likelihood (ML). However, differences between the
performances of the two scores in the earlier studies may be attributed to the
fact that they used approximate learning algorithms, not exact ones. This paper
compares the classification accuracies of BNs with approximate learning using
CLL to those with exact learning using ML. The results demonstrate that the
classification accuracies of BNs obtained by maximizing the ML are higher than
those obtained by maximizing the CLL for large data. However, the results also
demonstrate that the classification accuracies of exact learning BNs using the
ML are much worse than those of other methods when the sample size is small and
the class variable has numerous parents. To resolve the problem, we propose an
exact learning augmented naive Bayes classifier (ANB), which ensures a class
variable with no parents. The proposed method is guaranteed to asymptotically
estimate the identical class posterior to that of the exactly learned BN.
Comparison experiments demonstrated the superior performance of the proposed
method.Comment: 29 page