32,936 research outputs found
Phase transition in the sample complexity of likelihood-based phylogeny inference
Reconstructing evolutionary trees from molecular sequence data is a
fundamental problem in computational biology. Stochastic models of sequence
evolution are closely related to spin systems that have been extensively
studied in statistical physics and that connection has led to important
insights on the theoretical properties of phylogenetic reconstruction
algorithms as well as the development of new inference methods. Here, we study
maximum likelihood, a classical statistical technique which is perhaps the most
widely used in phylogenetic practice because of its superior empirical
accuracy.
At the theoretical level, except for its consistency, that is, the guarantee
of eventual correct reconstruction as the size of the input data grows, much
remains to be understood about the statistical properties of maximum likelihood
in this context. In particular, the best bounds on the sample complexity or
sequence-length requirement of maximum likelihood, that is, the amount of data
required for correct reconstruction, are exponential in the number, , of
tips---far from known lower bounds based on information-theoretic arguments.
Here we close the gap by proving a new upper bound on the sequence-length
requirement of maximum likelihood that matches up to constants the known lower
bound for some standard models of evolution.
More specifically, for the -state symmetric model of sequence evolution on
a binary phylogeny with bounded edge lengths, we show that the sequence-length
requirement behaves logarithmically in when the expected amount of mutation
per edge is below what is known as the Kesten-Stigum threshold. In general, the
sequence-length requirement is polynomial in . Our results imply moreover
that the maximum likelihood estimator can be computed efficiently on randomly
generated data provided sequences are as above.Comment: To appear in Probability Theory and Related Field
Reconstruction of Causal Networks by Set Covering
We present a method for the reconstruction of networks, based on the order of
nodes visited by a stochastic branching process. Our algorithm reconstructs a
network of minimal size that ensures consistency with the data. Crucially, we
show that global consistency with the data can be achieved through purely local
considerations, inferring the neighbourhood of each node in turn. The
optimisation problem solved for each individual node can be reduced to a Set
Covering Problem, which is known to be NP-hard but can be approximated well in
practice. We then extend our approach to account for noisy data, based on the
Minimum Description Length principle. We demonstrate our algorithms on
synthetic data, generated by an SIR-like epidemiological model.Comment: Under consideration for the ECML PKDD 2010 conferenc
- …