17 research outputs found
Hypergraph reconstruction from network data
Networks can describe the structure of a wide variety of complex systems by
specifying how pairs of nodes interact. This choice of representation is
flexible, but not necessarily appropriate when joint interactions between
groups of nodes are needed to explain empirical phenomena. Networks remain the
de facto standard, however, as relational datasets often fail to include
higher-order interactions. Here, we introduce a Bayesian approach to
reconstruct these missing higher-order interactions, from pairwise network
data. Our method is based on the principle of parsimony and only includes
higher-order structures when there is sufficient statistical evidence for them.Comment: 12 pages, 6 figures. Code is available at
https://graph-tool.skewed.de
Regularised inference for changepoint and dependency analysis in non-stationary processes
Multivariate correlated time series are found in many modern socio-scientific domains such as neurology, cyber-security, genetics and economics. The focus of this thesis is on efficiently modelling and inferring dependency structure both between data-streams and across points in time. In particular, it is considered that generating processes may vary over time, and are thus non-stationary. For example, patterns of brain activity are expected to change when performing different tasks or thought processes. Models that can describe such behaviour must be adaptable over time. However, such adaptability creates challenges for model identification. In order to perform learning or estimation one must control how model complexity grows in relation to the volume of data. To this extent, one of the main themes of this work is to investigate both the implementation and effect of assumptions on sparsity; relating to model parsimony at an individual time- point, and smoothness; how quickly a model may change over time. Throughout this thesis two basic classes of non-stationary model are stud- ied. Firstly, a class of piecewise constant Gaussian Graphical models (GGM) is introduced that can encode graphical dependencies between data-streams. In particular, a group-fused regulariser is examined that allows for the estima- tion of changepoints across graphical models. The second part of the thesis focuses on extending a class of locally-stationary wavelet (LSW) models. Un- like the raw GGM this enables one to encode dependencies not only between data-streams, but also across time. A set of sparsity aware estimators are developed for estimation of the spectral parameters of such models which are then compared to previous works in the domain
Applications of Probabilistic Inference to Planning & Reinforcement Learning
Optimal control is a profound and fascinating subject that regularly attracts interest from numerous scien-
tific disciplines, including both pure and applied Mathematics, Computer Science, Artificial Intelligence,
Psychology, Neuroscience and Economics. In
1960 Rudolf Kalman discovered that there exists a dual-
ity between the problems of filtering and optimal control in linear systems [84]. This is now regarded
as a seminal piece of work and it has since motivated a large amount of research into the discovery of
similar dualities between optimal control and statistical inference. This is especially true of recent years
where there has been much research into recasting problems of optimal control into problems of statis-
tical/approximate inference. Broadly speaking this is the perspective that we take in this work and in
particular we present various applications of methods from the fields of statistical/approximate inference
to optimal control, planning and Reinforcement Learning. Some of the methods would be more accu-
rately described to originate from other fields of research, such as the dual decomposition
techniques used in chapter(5) which originate from convex optimisation. However, the original motivation for the
application of these techniques was from the field of approximate inference. The study of dualities be-
tween optimal control and statistical inference has been a subject of research for over 50
years and we do not claim to encompass the entire subject. Instead, we present what we consider to be a range of
interesting and novel applications from this field of researc
On Competition for Undergraduate Co-op Placement: A Graph Approach
The objective of this thesis is to improve the co-operative (co-op) education process by analyzing the relationships among academic programs in the context of the co-op job market. To do this, we propose and apply a novel graph-mining methodology. The input to our problem consists of student-job interview pairs, with each student labelled with his or her academic program. From this input, we build a weighted directed graph, which we refer to as a program graph, in which nodes correspond to academic programs and edge weights denote the percentage of jobs that interviewed at least one student from both programs. For example, a directed edge from the Computer Engineering program to the Electrical Engineering program with weight 0.36 means that of all the jobs that interviewed at least one Computer Engineering student, 36 percent of those jobs also interviewed at least one Electrical Engineering student. Thus, the larger the edge weight, the stronger the relationship and competition between particular programs. The output consists of various graph properties and analyses, particularly those which find nodes forming clusters or communities, nodes that are connected to few or many clusters, and nodes that are strongly connected to their immediate neighbours. As we will show, these properties have natural interpretations in terms of the relationships among academic programs and competition for co-op jobs.
We applied the proposed methodology on one term of co-op interview data from a large Canadian university. We obtained interesting new insights that have not been reported in prior work. These insights can be beneficial to students, employers and academic institutions. Characterizing closely connected programs can help employers broaden their search for qualified students and can help students select programs of study that better correspond to their desired career. Students seeking a multi-disciplinary education can choose programs that are connected to other programs from many different clusters. Additionally, institutions can attend to programs that are strongly connected to (and face competition from) other programs by attracting more employers offering jobs in this area
Bayesian Gaussian Process Models: PAC-Bayesian Generalisation Error Bounds and Sparse Approximations
Institute for Adaptive and Neural ComputationNon-parametric models and techniques enjoy a growing popularity in the field of
machine learning, and among these Bayesian inference for Gaussian process (GP)
models has recently received significant attention. We feel that GP priors should
be part of the standard toolbox for constructing models relevant to machine
learning in the same way as parametric linear models are, and the results in this
thesis help to remove some obstacles on the way towards this goal.
In the first main chapter, we provide a distribution-free finite sample bound
on the difference between generalisation and empirical (training) error for GP
classification methods. While the general theorem (the PAC-Bayesian bound)
is not new, we give a much simplified and somewhat generalised derivation and
point out the underlying core technique (convex duality) explicitly. Furthermore,
the application to GP models is novel (to our knowledge). A central feature of
this bound is that its quality depends crucially on task knowledge being encoded
faithfully in the model and prior distributions, so there is a mutual benefit between
a sharp theoretical guarantee and empirically well-established statistical
practices. Extensive simulations on real-world classification tasks indicate an impressive
tightness of the bound, in spite of the fact that many previous bounds
for related kernel machines fail to give non-trivial guarantees in this practically
relevant regime.
In the second main chapter, sparse approximations are developed to address
the problem of the unfavourable scaling of most GP techniques with large training
sets. Due to its high importance in practice, this problem has received a lot of attention
recently. We demonstrate the tractability and usefulness of simple greedy
forward selection with information-theoretic criteria previously used in active
learning (or sequential design) and develop generic schemes for automatic model
selection with many (hyper)parameters. We suggest two new generic schemes and
evaluate some of their variants on large real-world classification and regression
tasks. These schemes and their underlying principles (which are clearly stated
and analysed) can be applied to obtain sparse approximations for a wide regime
of GP models far beyond the special cases we studied here
Blind image deconvolution: nonstationary Bayesian approaches to restoring blurred photos
High quality digital images have become pervasive in modern scientific and everyday life —
in areas from photography to astronomy, CCTV, microscopy, and medical imaging. However
there are always limits to the quality of these images due to uncertainty and imprecision in the
measurement systems. Modern signal processing methods offer the promise of overcoming
some of these problems by postprocessing
these blurred and noisy images. In this thesis,
novel methods using nonstationary statistical models are developed for the removal of blurs
from out of focus and other types of degraded photographic images.
The work tackles the fundamental problem blind image deconvolution (BID); its goal is
to restore a sharp image from a blurred observation when the blur itself is completely unknown.
This is a “doubly illposed”
problem — extreme lack of information must be countered
by strong prior constraints about sensible types of solution. In this work, the hierarchical
Bayesian methodology is used as a robust and versatile framework to impart the required prior
knowledge.
The thesis is arranged in two parts. In the first part, the BID problem is reviewed, along
with techniques and models for its solution. Observation models are developed, with an
emphasis on photographic restoration, concluding with a discussion of how these are reduced
to the common linear spatially-invariant
(LSI) convolutional model. Classical methods for the
solution of illposed
problems are summarised to provide a foundation for the main theoretical
ideas that will be used under the Bayesian framework. This is followed by an indepth
review
and discussion of the various prior image and blur models appearing in the literature, and then
their applications to solving the problem with both Bayesian and nonBayesian
techniques.
The second part covers novel restoration methods, making use of the theory presented in Part I.
Firstly, two new nonstationary image models are presented. The first models local variance in
the image, and the second extends this with locally adaptive noncausal
autoregressive (AR)
texture estimation and local mean components. These models allow for recovery of image
details including edges and texture, whilst preserving smooth regions. Most existing methods
do not model the boundary conditions correctly for deblurring of natural photographs, and a
Chapter is devoted to exploring Bayesian solutions to this topic.
Due to the complexity of the models used and the problem itself, there are many challenges
which must be overcome for tractable inference. Using the new models, three different inference
strategies are investigated: firstly using the Bayesian maximum marginalised a posteriori
(MMAP) method with deterministic optimisation; proceeding with the stochastic methods
of variational Bayesian (VB) distribution approximation, and simulation of the posterior distribution
using the Gibbs sampler. Of these, we find the Gibbs sampler to be the most effective
way to deal with a variety of different types of unknown blurs. Along the way, details are given
of the numerical strategies developed to give accurate results and to accelerate performance.
Finally, the thesis demonstrates state of the art
results in blind restoration of synthetic and real
degraded images, such as recovering details in out of focus photographs