12 research outputs found
Graph construction with condition-based weights for spectral clustering of hierarchical datasets
Most of the unsupervised machine learning algorithms focus on clustering the data based on similarity metrics, while ignoring other attributes, or perhaps other type of connections between the data points. In case of hierarchical datasets, groups of points (point-sets) can be defined according to the hierarchy system. Our goal was to develop such spectral clustering approach that preserves the structure of the dataset throughout the clustering procedure. The main contribution of this paper is a set of conditions for weighted graph construction used in spectral clustering. Following the requirements – given by the set of conditions – ensures that the hierarchical formation of the dataset remains unchanged, and therefore the clustering of data points imply the clustering of point-sets as well. The proposed spectral clustering algorithm was tested on three datasets, the results were compared to baseline methods and it can be concluded the algorithm with the proposed conditions always preserves the hierarchy structure
A Variational View on Bootstrap Ensembles as Bayesian Inference
In this paper, we employ variational arguments to establish a connection
between ensemble methods for Neural Networks and Bayesian inference. We
consider an ensemble-based scheme where each model/particle corresponds to a
perturbation of the data by means of parametric bootstrap and a perturbation of
the prior. We derive conditions under which any optimization steps of the
particles makes the associated distribution reduce its divergence to the
posterior over model parameters. Such conditions do not require any particular
form for the approximation and they are purely geometrical, giving insights on
the behavior of the ensemble on a number of interesting models such as Neural
Networks with ReLU activations. Experiments confirm that ensemble methods can
be a valid alternative to approximate Bayesian inference; the theoretical
developments in the paper seek to explain this behavior
Sparse Gaussian Processes Revisited: Bayesian Approaches to Inducing-Variable Approximations
Variational inference techniques based on inducing variables provide an
elegant framework for scalable posterior estimation in Gaussian process (GP)
models. Besides enabling scalability, one of their main advantages over sparse
approximations using direct marginal likelihood maximization is that they
provide a robust alternative for point estimation of the inducing inputs, i.e.
the location of the inducing variables. In this work we challenge the common
wisdom that optimizing the inducing inputs in the variational framework yields
optimal performance. We show that, by revisiting old model approximations such
as the fully-independent training conditionals endowed with powerful
sampling-based inference methods, treating both inducing locations and GP
hyper-parameters in a Bayesian way can improve performance significantly. Based
on stochastic gradient Hamiltonian Monte Carlo, we develop a fully Bayesian
approach to scalable GP and deep GP models, and demonstrate its
state-of-the-art performance through an extensive experimental campaign across
several regression and classification problems