10 research outputs found
VaiPhy: a Variational Inference Based Algorithm for Phylogeny
Phylogenetics is a classical methodology in computational biology that today
has become highly relevant for medical investigation of single-cell data, e.g.,
in the context of cancer development. The exponential size of the tree space
is, unfortunately, a substantial obstacle for Bayesian phylogenetic inference
using Markov chain Monte Carlo based methods since these rely on local
operations. And although more recent variational inference (VI) based methods
offer speed improvements, they rely on expensive auto-differentiation
operations for learning the variational parameters. We propose VaiPhy, a
remarkably fast VI based algorithm for approximate posterior inference in an
augmented tree space. VaiPhy produces marginal log-likelihood estimates on par
with the state-of-the-art methods on real data and is considerably faster since
it does not require auto-differentiation. Instead, VaiPhy combines coordinate
ascent update equations with two novel sampling schemes: (i) SLANTIS, a
proposal distribution for tree topologies in the augmented tree space, and (ii)
the JC sampler, to the best of our knowledge, the first-ever scheme for
sampling branch lengths directly from the popular Jukes-Cantor model. We
compare VaiPhy in terms of density estimation and runtime. Additionally, we
evaluate the reproducibility of the baselines. We provide our code on GitHub:
\url{https://github.com/Lagergren-Lab/VaiPhy}.Comment: NeurIPS-22 conference pape
Learning with MISELBO: The Mixture Cookbook
Mixture models in variational inference (VI) is an active field of research.
Recent works have established their connection to multiple importance sampling
(MIS) through the MISELBO and advanced the use of ensemble approximations for
large-scale problems. However, as we show here, an independent learning of the
ensemble components can lead to suboptimal diversity. Hence, we study the
effect of instead using MISELBO as an objective function for learning mixtures,
and we propose the first ever mixture of variational approximations for a
normalizing flow-based hierarchical variational autoencoder (VAE) with
VampPrior and a PixelCNN decoder network. Two major insights led to the
construction of this novel composite model. First, mixture models have
potential to be off-the-shelf tools for practitioners to obtain more flexible
posterior approximations in VAEs. Therefore, we make them more accessible by
demonstrating how to apply them to four popular architectures. Second, the
mixture components cooperate in order to cover the target distribution while
trying to maximize their diversity when MISELBO is the objective function. We
explain this cooperative behavior by drawing a novel connection between VI and
adaptive importance sampling. Finally, we demonstrate the superiority of the
Mixture VAEs' learned feature representations on both image and single-cell
transcriptome data, and obtain state-of-the-art results among VAE architectures
in terms of negative log-likelihood on the MNIST and FashionMNIST datasets.
Code available here: \url{https://github.com/Lagergren-Lab/MixtureVAEs}
KL/TV Reshuffling : Statistical Distance Based Offspring Selection in SMC Methods
Over the years sequential Monte Carlo (SMC), and, equivalently, particle filter (PF) theory has enjoyed much attention from researchers. However, the intensity of developing innovative resampling methods, also known as offspring selection methods, has long been declining, with most of the popular schemes aging back two decades. Especially, the set of deterministic offspring selection methods is limited. In light of this, and inspired by variational inference, we propose offspring selection schemes which multiply/discard particles in order to minimize statistical distances between relevant distributions. By regarding offspring selection as a problem of minimizing statistical distances, we further bridge the gap between optimisation-based density estimation and SMC theory. Our contribution is in a sense twofold. Partly, we provide novel, deterministic offspring selection schemes, and, partly, we extend the class of SMC algorithms by using the particle likelihoods instead of importance weights when doing offspring selection. Our proposed methods outperform or compare favourably with the two most popular resampling schemes on density-estimation benchmark tests, which are commonly turned to in the SMC and particle Markov chain Monte Carlo (PMCMC) literature.  Under Ären har teorin inom sekventiell Monte Carlo (SMC) och, likvÀl, partikelfilter (PF) fÄtt stor uppmÀrksamhet frÄn forskare. Intensiteten att utveckla innovativa metoder för urval av avkommor, har dock lÀnge avtagit och de flesta av de populÀra systemen har funnits i tvÄ decennier. Speciellt Àr uppsÀttningen av deterministiska urvalsmetoder begrÀnsad. Mot bakgrund av detta, och inspirerad av variationsslutledning, föreslÄr vi urvalsmetoder som multiplicerar / kasserar partiklar för att minimera statistiska avstÄnd mellan relevanta fördelningar. Genom att betrakta urvalet som ett problem för att minimera statistiska avstÄnd, överbryggar vi ytterligare klyftan mellan optimeringsbaserad densitetsuppskattning och SMC-teori. VÄrt bidrag Àr pÄ sÀtt och vis dubbelt. Delvis tillhandahÄller vi nya, deterministiska urvalsscheman, och delvis utökar vi klassen av SMC-algoritmer genom att anvÀnda partikel sannolikheter istÀllet för viktvikter nÀr man gör avkommesval. VÄra föreslagna metoder övertrÀffar eller jÀmför fördelaktigt med de tvÄ mest populÀra urvalsmetoderna för densitetsuppskattning pÄ test som vanligtvis anvÀnds för utvÀrdera metoder inom SMC och Markov-kedje-Monte Carlo
KL/TV Reshuffling : Statistical Distance Based Offspring Selection in SMC Methods
Over the years sequential Monte Carlo (SMC), and, equivalently, particle filter (PF) theory has enjoyed much attention from researchers. However, the intensity of developing innovative resampling methods, also known as offspring selection methods, has long been declining, with most of the popular schemes aging back two decades. Especially, the set of deterministic offspring selection methods is limited. In light of this, and inspired by variational inference, we propose offspring selection schemes which multiply/discard particles in order to minimize statistical distances between relevant distributions. By regarding offspring selection as a problem of minimizing statistical distances, we further bridge the gap between optimisation-based density estimation and SMC theory. Our contribution is in a sense twofold. Partly, we provide novel, deterministic offspring selection schemes, and, partly, we extend the class of SMC algorithms by using the particle likelihoods instead of importance weights when doing offspring selection. Our proposed methods outperform or compare favourably with the two most popular resampling schemes on density-estimation benchmark tests, which are commonly turned to in the SMC and particle Markov chain Monte Carlo (PMCMC) literature.  Under Ären har teorin inom sekventiell Monte Carlo (SMC) och, likvÀl, partikelfilter (PF) fÄtt stor uppmÀrksamhet frÄn forskare. Intensiteten att utveckla innovativa metoder för urval av avkommor, har dock lÀnge avtagit och de flesta av de populÀra systemen har funnits i tvÄ decennier. Speciellt Àr uppsÀttningen av deterministiska urvalsmetoder begrÀnsad. Mot bakgrund av detta, och inspirerad av variationsslutledning, föreslÄr vi urvalsmetoder som multiplicerar / kasserar partiklar för att minimera statistiska avstÄnd mellan relevanta fördelningar. Genom att betrakta urvalet som ett problem för att minimera statistiska avstÄnd, överbryggar vi ytterligare klyftan mellan optimeringsbaserad densitetsuppskattning och SMC-teori. VÄrt bidrag Àr pÄ sÀtt och vis dubbelt. Delvis tillhandahÄller vi nya, deterministiska urvalsscheman, och delvis utökar vi klassen av SMC-algoritmer genom att anvÀnda partikel sannolikheter istÀllet för viktvikter nÀr man gör avkommesval. VÄra föreslagna metoder övertrÀffar eller jÀmför fördelaktigt med de tvÄ mest populÀra urvalsmetoderna för densitetsuppskattning pÄ test som vanligtvis anvÀnds för utvÀrdera metoder inom SMC och Markov-kedje-Monte Carlo
Applicability of a Translucent Barrier Based Model of Noise
The aim of this project was to create our own data set consisting of images of fruits and vegetables. A subset of the data set was composed of images where the fruits and vegetables were obscured by a plastic bag. We then evaluated the difficulty of this data set using a simple kernel machine algorithm. The performance drops considerably when introducing the above mentioned subset to the data set. The algorithm was to classify the different types of fruits and vegetables present in the data set. We also created the data set in different pixel dimensions, sufficiently reducing the computation time of the algorithm while not suffering a large drop in classification performance. This enables algorithms which complexity are highly dependent on input dimension size to use the data set. From our different experimental setups we were able to conclude that the machine outperforms humans on small input dimensions, given that the humans had no prior knowledge of the data set
Applicability of a Translucent Barrier Based Model of Noise
The aim of this project was to create our own data set consisting of images of fruits and vegetables. A subset of the data set was composed of images where the fruits and vegetables were obscured by a plastic bag. We then evaluated the difficulty of this data set using a simple kernel machine algorithm. The performance drops considerably when introducing the above mentioned subset to the data set. The algorithm was to classify the different types of fruits and vegetables present in the data set. We also created the data set in different pixel dimensions, sufficiently reducing the computation time of the algorithm while not suffering a large drop in classification performance. This enables algorithms which complexity are highly dependent on input dimension size to use the data set. From our different experimental setups we were able to conclude that the machine outperforms humans on small input dimensions, given that the humans had no prior knowledge of the data set
Applicability of a Translucent Barrier Based Model of Noise
The aim of this project was to create our own data set consisting of images of fruits and vegetables. A subset of the data set was composed of images where the fruits and vegetables were obscured by a plastic bag. We then evaluated the difficulty of this data set using a simple kernel machine algorithm. The performance drops considerably when introducing the above mentioned subset to the data set. The algorithm was to classify the different types of fruits and vegetables present in the data set. We also created the data set in different pixel dimensions, sufficiently reducing the computation time of the algorithm while not suffering a large drop in classification performance. This enables algorithms which complexity are highly dependent on input dimension size to use the data set. From our different experimental setups we were able to conclude that the machine outperforms humans on small input dimensions, given that the humans had no prior knowledge of the data set
Multiple Importance Sampling ELBO and Deep Ensembles of Variational Approximations
In variational inference (VI), the marginal log-likelihood is estimated using
the standard evidence lower bound (ELBO), or improved versions as the
importance weighted ELBO (IWELBO). We propose the multiple importance sampling
ELBO (MISELBO), a \textit{versatile} yet \textit{simple} framework. MISELBO is
applicable in both amortized and classical VI, and it uses ensembles, e.g.,
deep ensembles, of independently inferred variational approximations. As far as
we are aware, the concept of deep ensembles in amortized VI has not previously
been established. We prove that MISELBO provides a tighter bound than the
average of standard ELBOs, and demonstrate empirically that it gives tighter
bounds than the average of IWELBOs. MISELBO is evaluated in density-estimation
experiments that include MNIST and several real-data phylogenetic tree
inference problems. First, on the MNIST dataset, MISELBO boosts the
density-estimation performances of a state-of-the-art model, nouveau VAE.
Second, in the phylogenetic tree inference setting, our framework enhances a
state-of-the-art VI algorithm that uses normalizing flows. On top of the
technical benefits of MISELBO, it allows to unveil connections between VI and
recent advances in the importance sampling literature, paving the way for
further methodological advances. We provide our code at
\url{https://github.com/Lagergren-Lab/MISELBO}.Comment: AISTATS 202
VaiPhy: a Variational Inference Based Algorithm for Phylogeny
Phylogenetics is a classical methodology in com- putational biology that today has become highly relevant for medical investigation of single-cell data, e.g., in the context of development of can- cer. The exponential size of the tree space is unfortunately a formidable obstacle for current Bayesian phylogenetic inference using Markov chain Monte Carlo based methods since these rely on local operations. And although more re- cent variational inference (VI) based methods of- fer speed improvements, they rely on expensive auto-differentiation operations for learning the variational parameters. We propose VaiPhy, a remarkably fast VI based algorithm for approx- imate posterior inference in an augmented tree space. VaiPhy produces marginal log-likelihood estimates on par with the state-of-the-art meth- ods on real data, and is considerably faster since it does not require auto-differentiation. Instead, VaiPhy combines coordinate ascent update equa- tions with two novel sampling schemes: (i) SLANTIS, a proposal distribution for tree topolo- gies in the augmented tree space, and (ii) the JC sampler, the, to the best of our knowledge, first ever scheme for sampling branch lengths directly from the popular Jukes-Cantor model. We compare VaiPhy in terms of density esti- mation and runtime. Additionally, we evaluate the reproducibility of the baselines. We provide our code on GitHub: https://github.com/ Lagergren-Lab/VaiPhy. QC 20220421</p
VaiPhy: a Variational Inference Based Algorithm for Phylogeny
Phylogenetics is a classical methodology in com- putational biology that today has become highly relevant for medical investigation of single-cell data, e.g., in the context of development of can- cer. The exponential size of the tree space is unfortunately a formidable obstacle for current Bayesian phylogenetic inference using Markov chain Monte Carlo based methods since these rely on local operations. And although more re- cent variational inference (VI) based methods of- fer speed improvements, they rely on expensive auto-differentiation operations for learning the variational parameters. We propose VaiPhy, a remarkably fast VI based algorithm for approx- imate posterior inference in an augmented tree space. VaiPhy produces marginal log-likelihood estimates on par with the state-of-the-art meth- ods on real data, and is considerably faster since it does not require auto-differentiation. Instead, VaiPhy combines coordinate ascent update equa- tions with two novel sampling schemes: (i) SLANTIS, a proposal distribution for tree topolo- gies in the augmented tree space, and (ii) the JC sampler, the, to the best of our knowledge, first ever scheme for sampling branch lengths directly from the popular Jukes-Cantor model. We compare VaiPhy in terms of density esti- mation and runtime. Additionally, we evaluate the reproducibility of the baselines. We provide our code on GitHub: https://github.com/ Lagergren-Lab/VaiPhy. QC 20220421</p