1,913 research outputs found
Non-informative reparameterisations for location-scale mixtures
While mixtures of Gaussian distributions have been studied for more than a century (Pearson, 1894), the construction of a reference Bayesian analysis of those models still remains unsolved, with a general prohibition of the usage of improper priors (Fr\"uwirth-Schnatter, 2006) due to the ill-posed nature of such statistical objects. This difficulty is usually bypassed by an empirical Bayes resolution (Richardson and Green, 1997). By creating a new parameterisation cantered on the mean and variance of the mixture distribution itself, we are able to develop here a genuine non-informative prior for Gaussian mixtures with an arbitrary number of components. We demonstrate that the posterior distribution associated with this prior is almost surely proper and provide MCMC implementations that exhibit the expected exchangeability. While we only study here the Gaussian case, extension to other classes of location-scale mixtures is straightforward
A Quantile Variant of the EM Algorithm and Its Applications to Parameter Estimation with Interval Data
The expectation-maximization (EM) algorithm is a powerful computational
technique for finding the maximum likelihood estimates for parametric models
when the data are not fully observed. The EM is best suited for situations
where the expectation in each E-step and the maximization in each M-step are
straightforward. A difficulty with the implementation of the EM algorithm is
that each E-step requires the integration of the log-likelihood function in
closed form. The explicit integration can be avoided by using what is known as
the Monte Carlo EM (MCEM) algorithm. The MCEM uses a random sample to estimate
the integral at each E-step. However, the problem with the MCEM is that it
often converges to the integral quite slowly and the convergence behavior can
also be unstable, which causes a computational burden. In this paper, we
propose what we refer to as the quantile variant of the EM (QEM) algorithm. We
prove that the proposed QEM method has an accuracy of while the MCEM
method has an accuracy of . Thus, the proposed QEM method
possesses faster and more stable convergence properties when compared with the
MCEM algorithm. The improved performance is illustrated through the numerical
studies. Several practical examples illustrating its use in interval-censored
data problems are also provided
Shear Modes, Criticality and Extremal Black Holes
We consider a (2+1)-dimensional field theory, assumed to be holographically
dual to the extremal Reissner-Nordstrom AdS(4) black hole background, and
calculate the retarded correlators of charge (vector) current and
energy-momentum (tensor) operators at finite momentum and frequency. We show
that, similar to what was observed previously for the correlators of scalar and
spinor operators, these correlators exhibit emergent scaling behavior at low
frequency. We numerically compute the electromagnetic and gravitational
quasinormal frequencies (in the shear channel) of the extremal
Reissner-Nordstrom AdS(4) black hole corresponding to the spectrum of poles in
the retarded correlators. The picture that emerges is quite simple: there is a
branch cut along the negative imaginary frequency axis, and a series of
isolated poles corresponding to damped excitations. All of these poles are
always in the lower half complex frequency plane, indicating stability. We show
that this analytic structure can be understood as the proper limit of finite
temperature results as T is taken to zero holding the chemical potential fixed.Comment: 28 pages, 7 figures, added reference
Sets of Priors Reflecting Prior-Data Conflict and Agreement
In Bayesian statistics, the choice of prior distribution is often debatable,
especially if prior knowledge is limited or data are scarce. In imprecise
probability, sets of priors are used to accurately model and reflect prior
knowledge. This has the advantage that prior-data conflict sensitivity can be
modelled: Ranges of posterior inferences should be larger when prior and data
are in conflict. We propose a new method for generating prior sets which, in
addition to prior-data conflict sensitivity, allows to reflect strong
prior-data agreement by decreased posterior imprecision.Comment: 12 pages, 6 figures, In: Paulo Joao Carvalho et al. (eds.), IPMU
2016: Proceedings of the 16th International Conference on Information
Processing and Management of Uncertainty in Knowledge-Based Systems,
Eindhoven, The Netherland
Assessing the Potential of Classical Q-learning in General Game Playing
After the recent groundbreaking results of AlphaGo and AlphaZero, we have
seen strong interests in deep reinforcement learning and artificial general
intelligence (AGI) in game playing. However, deep learning is
resource-intensive and the theory is not yet well developed. For small games,
simple classical table-based Q-learning might still be the algorithm of choice.
General Game Playing (GGP) provides a good testbed for reinforcement learning
to research AGI. Q-learning is one of the canonical reinforcement learning
methods, and has been used by (Banerjee Stone, IJCAI 2007) in GGP. In this
paper we implement Q-learning in GGP for three small-board games (Tic-Tac-Toe,
Connect Four, Hex)\footnote{source code: https://github.com/wh1992v/ggp-rl}, to
allow comparison to Banerjee et al.. We find that Q-learning converges to a
high win rate in GGP. For the -greedy strategy, we propose a first
enhancement, the dynamic algorithm. In addition, inspired by (Gelly
Silver, ICML 2007) we combine online search (Monte Carlo Search) to
enhance offline learning, and propose QM-learning for GGP. Both enhancements
improve the performance of classical Q-learning. In this work, GGP allows us to
show, if augmented by appropriate enhancements, that classical table-based
Q-learning can perform well in small games.Comment: arXiv admin note: substantial text overlap with arXiv:1802.0594
The binaural masking level difference: cortical correlates persist despite severe brain stem atrophy in progressive supranuclear palsy.
Under binaural listening conditions, the detection of target signals within background masking noise is substantially improved when the interaural phase of the target differs from that of the masker. Neural correlates of this binaural masking level difference (BMLD) have been observed in the inferior colliculus and temporal cortex, but it is not known whether degeneration of the inferior colliculus would result in a reduction of the BMLD in humans. We used magnetoencephalography to examine the BMLD in 13 healthy adults and 13 patients with progressive supranuclear palsy (PSP). PSP is associated with severe atrophy of the upper brain stem, including the inferior colliculus, confirmed by voxel-based morphometry of structural MRI. Stimuli comprised in-phase sinusoidal tones presented to both ears at three levels (high, medium, and low) masked by in-phase noise, which rendered the low-level tone inaudible. Critically, the BMLD was measured using a low-level tone presented in opposite phase across ears, making it audible against the noise. The cortical waveforms from bilateral auditory sources revealed significantly larger N1m peaks for the out-of-phase low-level tone compared with the in-phase low-level tone, for both groups, indicating preservation of early cortical correlates of the BMLD in PSP. In PSP a significant delay was observed in the onset of the N1m deflection and the amplitude of the P2m was reduced, but these differences were not restricted to the BMLD condition. The results demonstrate that although PSP causes subtle auditory deficits, binaural processing can survive the presence of significant damage to the upper brain stem.This work has been supported by the Wellcome Trust (Grants 088324 and 088263); Medical Research Council (G0700503 to B. C. P. Ghosh); Guarantors of Brain (to B. C. P. Ghosh); Raymond and Beverley Sackler Trust (to B. C. P. Ghosh); and National Institute of Health Research Cambridge Comprehensive Biomedical Research Centre including the CambridgeBrain Bank.This is the final version of the article. It first appeared from American Physiological Society via http://dx.doi.org/10.1152/jn.00062.201
Probabilistic Clustering of Time-Evolving Distance Data
We present a novel probabilistic clustering model for objects that are
represented via pairwise distances and observed at different time points. The
proposed method utilizes the information given by adjacent time points to find
the underlying cluster structure and obtain a smooth cluster evolution. This
approach allows the number of objects and clusters to differ at every time
point, and no identification on the identities of the objects is needed.
Further, the model does not require the number of clusters being specified in
advance -- they are instead determined automatically using a Dirichlet process
prior. We validate our model on synthetic data showing that the proposed method
is more accurate than state-of-the-art clustering methods. Finally, we use our
dynamic clustering model to analyze and illustrate the evolution of brain
cancer patients over time
Synthesizing and tuning chemical reaction networks with specified behaviours
We consider how to generate chemical reaction networks (CRNs) from functional
specifications. We propose a two-stage approach that combines synthesis by
satisfiability modulo theories and Markov chain Monte Carlo based optimisation.
First, we identify candidate CRNs that have the possibility to produce correct
computations for a given finite set of inputs. We then optimise the reaction
rates of each CRN using a combination of stochastic search techniques applied
to the chemical master equation, simultaneously improving the of correct
behaviour and ruling out spurious solutions. In addition, we use techniques
from continuous time Markov chain theory to study the expected termination time
for each CRN. We illustrate our approach by identifying CRNs for majority
decision-making and division computation, which includes the identification of
both known and unknown networks.Comment: 17 pages, 6 figures, appeared the proceedings of the 21st conference
on DNA Computing and Molecular Programming, 201
Coupled coarse graining and Markov Chain Monte Carlo for lattice systems
We propose an efficient Markov Chain Monte Carlo method for sampling
equilibrium distributions for stochastic lattice models, capable of handling
correctly long and short-range particle interactions. The proposed method is a
Metropolis-type algorithm with the proposal probability transition matrix based
on the coarse-grained approximating measures introduced in a series of works of
M. Katsoulakis, A. Majda, D. Vlachos and P. Plechac, L. Rey-Bellet and
D.Tsagkarogiannis,. We prove that the proposed algorithm reduces the
computational cost due to energy differences and has comparable mixing
properties with the classical microscopic Metropolis algorithm, controlled by
the level of coarsening and reconstruction procedure. The properties and
effectiveness of the algorithm are demonstrated with an exactly solvable
example of a one dimensional Ising-type model, comparing efficiency of the
single spin-flip Metropolis dynamics and the proposed coupled Metropolis
algorithm.Comment: 20 pages, 4 figure
Fermions and Type IIB Supergravity On Squashed Sasaki-Einstein Manifolds
We discuss the dimensional reduction of fermionic modes in a recently found
class of consistent truncations of type IIB supergravity compactified on
squashed five-dimensional Sasaki-Einstein manifolds. We derive the lower
dimensional equations of motion and effective action, and comment on the
supersymmetry of the resulting theory, which is consistent with N=4 gauged
supergravity in , coupled to two vector multiplets. We compute fermion
masses by linearizing around two vacua of the theory: one that breaks
N=4 down to N=2 spontaneously, and a second one which preserves no
supersymmetries. The truncations under consideration are noteworthy in that
they retain massive modes which are charged under a U(1) subgroup of the
-symmetry, a feature that makes them interesting for applications to
condensed matter phenomena via gauge/gravity duality. In this light, as an
application of our general results we exhibit the coupling of the fermions to
the type IIB holographic superconductor, and find a consistent further
truncation of the fermion sector that retains a single spin-1/2 mode.Comment: 43 pages, 2 figures, PDFLaTeX; v2: added references, typos corrected,
minor change
- …
