1,913 research outputs found

    Non-informative reparameterisations for location-scale mixtures

    Get PDF
    While mixtures of Gaussian distributions have been studied for more than a century (Pearson, 1894), the construction of a reference Bayesian analysis of those models still remains unsolved, with a general prohibition of the usage of improper priors (Fr\"uwirth-Schnatter, 2006) due to the ill-posed nature of such statistical objects. This difficulty is usually bypassed by an empirical Bayes resolution (Richardson and Green, 1997). By creating a new parameterisation cantered on the mean and variance of the mixture distribution itself, we are able to develop here a genuine non-informative prior for Gaussian mixtures with an arbitrary number of components. We demonstrate that the posterior distribution associated with this prior is almost surely proper and provide MCMC implementations that exhibit the expected exchangeability. While we only study here the Gaussian case, extension to other classes of location-scale mixtures is straightforward

    A Quantile Variant of the EM Algorithm and Its Applications to Parameter Estimation with Interval Data

    Full text link
    The expectation-maximization (EM) algorithm is a powerful computational technique for finding the maximum likelihood estimates for parametric models when the data are not fully observed. The EM is best suited for situations where the expectation in each E-step and the maximization in each M-step are straightforward. A difficulty with the implementation of the EM algorithm is that each E-step requires the integration of the log-likelihood function in closed form. The explicit integration can be avoided by using what is known as the Monte Carlo EM (MCEM) algorithm. The MCEM uses a random sample to estimate the integral at each E-step. However, the problem with the MCEM is that it often converges to the integral quite slowly and the convergence behavior can also be unstable, which causes a computational burden. In this paper, we propose what we refer to as the quantile variant of the EM (QEM) algorithm. We prove that the proposed QEM method has an accuracy of O(1/K2)O(1/K^2) while the MCEM method has an accuracy of Op(1/K)O_p(1/\sqrt{K}). Thus, the proposed QEM method possesses faster and more stable convergence properties when compared with the MCEM algorithm. The improved performance is illustrated through the numerical studies. Several practical examples illustrating its use in interval-censored data problems are also provided

    Shear Modes, Criticality and Extremal Black Holes

    Full text link
    We consider a (2+1)-dimensional field theory, assumed to be holographically dual to the extremal Reissner-Nordstrom AdS(4) black hole background, and calculate the retarded correlators of charge (vector) current and energy-momentum (tensor) operators at finite momentum and frequency. We show that, similar to what was observed previously for the correlators of scalar and spinor operators, these correlators exhibit emergent scaling behavior at low frequency. We numerically compute the electromagnetic and gravitational quasinormal frequencies (in the shear channel) of the extremal Reissner-Nordstrom AdS(4) black hole corresponding to the spectrum of poles in the retarded correlators. The picture that emerges is quite simple: there is a branch cut along the negative imaginary frequency axis, and a series of isolated poles corresponding to damped excitations. All of these poles are always in the lower half complex frequency plane, indicating stability. We show that this analytic structure can be understood as the proper limit of finite temperature results as T is taken to zero holding the chemical potential fixed.Comment: 28 pages, 7 figures, added reference

    Sets of Priors Reflecting Prior-Data Conflict and Agreement

    Full text link
    In Bayesian statistics, the choice of prior distribution is often debatable, especially if prior knowledge is limited or data are scarce. In imprecise probability, sets of priors are used to accurately model and reflect prior knowledge. This has the advantage that prior-data conflict sensitivity can be modelled: Ranges of posterior inferences should be larger when prior and data are in conflict. We propose a new method for generating prior sets which, in addition to prior-data conflict sensitivity, allows to reflect strong prior-data agreement by decreased posterior imprecision.Comment: 12 pages, 6 figures, In: Paulo Joao Carvalho et al. (eds.), IPMU 2016: Proceedings of the 16th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, Eindhoven, The Netherland

    Assessing the Potential of Classical Q-learning in General Game Playing

    Get PDF
    After the recent groundbreaking results of AlphaGo and AlphaZero, we have seen strong interests in deep reinforcement learning and artificial general intelligence (AGI) in game playing. However, deep learning is resource-intensive and the theory is not yet well developed. For small games, simple classical table-based Q-learning might still be the algorithm of choice. General Game Playing (GGP) provides a good testbed for reinforcement learning to research AGI. Q-learning is one of the canonical reinforcement learning methods, and has been used by (Banerjee &\& Stone, IJCAI 2007) in GGP. In this paper we implement Q-learning in GGP for three small-board games (Tic-Tac-Toe, Connect Four, Hex)\footnote{source code: https://github.com/wh1992v/ggp-rl}, to allow comparison to Banerjee et al.. We find that Q-learning converges to a high win rate in GGP. For the ϵ\epsilon-greedy strategy, we propose a first enhancement, the dynamic ϵ\epsilon algorithm. In addition, inspired by (Gelly &\& Silver, ICML 2007) we combine online search (Monte Carlo Search) to enhance offline learning, and propose QM-learning for GGP. Both enhancements improve the performance of classical Q-learning. In this work, GGP allows us to show, if augmented by appropriate enhancements, that classical table-based Q-learning can perform well in small games.Comment: arXiv admin note: substantial text overlap with arXiv:1802.0594

    The binaural masking level difference: cortical correlates persist despite severe brain stem atrophy in progressive supranuclear palsy.

    Get PDF
    Under binaural listening conditions, the detection of target signals within background masking noise is substantially improved when the interaural phase of the target differs from that of the masker. Neural correlates of this binaural masking level difference (BMLD) have been observed in the inferior colliculus and temporal cortex, but it is not known whether degeneration of the inferior colliculus would result in a reduction of the BMLD in humans. We used magnetoencephalography to examine the BMLD in 13 healthy adults and 13 patients with progressive supranuclear palsy (PSP). PSP is associated with severe atrophy of the upper brain stem, including the inferior colliculus, confirmed by voxel-based morphometry of structural MRI. Stimuli comprised in-phase sinusoidal tones presented to both ears at three levels (high, medium, and low) masked by in-phase noise, which rendered the low-level tone inaudible. Critically, the BMLD was measured using a low-level tone presented in opposite phase across ears, making it audible against the noise. The cortical waveforms from bilateral auditory sources revealed significantly larger N1m peaks for the out-of-phase low-level tone compared with the in-phase low-level tone, for both groups, indicating preservation of early cortical correlates of the BMLD in PSP. In PSP a significant delay was observed in the onset of the N1m deflection and the amplitude of the P2m was reduced, but these differences were not restricted to the BMLD condition. The results demonstrate that although PSP causes subtle auditory deficits, binaural processing can survive the presence of significant damage to the upper brain stem.This work has been supported by the Wellcome Trust (Grants 088324 and 088263); Medical Research Council (G0700503 to B. C. P. Ghosh); Guarantors of Brain (to B. C. P. Ghosh); Raymond and Beverley Sackler Trust (to B. C. P. Ghosh); and National Institute of Health Research Cambridge Comprehensive Biomedical Research Centre including the CambridgeBrain Bank.This is the final version of the article. It first appeared from American Physiological Society via http://dx.doi.org/10.1152/jn.00062.201

    Probabilistic Clustering of Time-Evolving Distance Data

    Full text link
    We present a novel probabilistic clustering model for objects that are represented via pairwise distances and observed at different time points. The proposed method utilizes the information given by adjacent time points to find the underlying cluster structure and obtain a smooth cluster evolution. This approach allows the number of objects and clusters to differ at every time point, and no identification on the identities of the objects is needed. Further, the model does not require the number of clusters being specified in advance -- they are instead determined automatically using a Dirichlet process prior. We validate our model on synthetic data showing that the proposed method is more accurate than state-of-the-art clustering methods. Finally, we use our dynamic clustering model to analyze and illustrate the evolution of brain cancer patients over time

    Synthesizing and tuning chemical reaction networks with specified behaviours

    Full text link
    We consider how to generate chemical reaction networks (CRNs) from functional specifications. We propose a two-stage approach that combines synthesis by satisfiability modulo theories and Markov chain Monte Carlo based optimisation. First, we identify candidate CRNs that have the possibility to produce correct computations for a given finite set of inputs. We then optimise the reaction rates of each CRN using a combination of stochastic search techniques applied to the chemical master equation, simultaneously improving the of correct behaviour and ruling out spurious solutions. In addition, we use techniques from continuous time Markov chain theory to study the expected termination time for each CRN. We illustrate our approach by identifying CRNs for majority decision-making and division computation, which includes the identification of both known and unknown networks.Comment: 17 pages, 6 figures, appeared the proceedings of the 21st conference on DNA Computing and Molecular Programming, 201

    Coupled coarse graining and Markov Chain Monte Carlo for lattice systems

    Get PDF
    We propose an efficient Markov Chain Monte Carlo method for sampling equilibrium distributions for stochastic lattice models, capable of handling correctly long and short-range particle interactions. The proposed method is a Metropolis-type algorithm with the proposal probability transition matrix based on the coarse-grained approximating measures introduced in a series of works of M. Katsoulakis, A. Majda, D. Vlachos and P. Plechac, L. Rey-Bellet and D.Tsagkarogiannis,. We prove that the proposed algorithm reduces the computational cost due to energy differences and has comparable mixing properties with the classical microscopic Metropolis algorithm, controlled by the level of coarsening and reconstruction procedure. The properties and effectiveness of the algorithm are demonstrated with an exactly solvable example of a one dimensional Ising-type model, comparing efficiency of the single spin-flip Metropolis dynamics and the proposed coupled Metropolis algorithm.Comment: 20 pages, 4 figure

    Fermions and Type IIB Supergravity On Squashed Sasaki-Einstein Manifolds

    Full text link
    We discuss the dimensional reduction of fermionic modes in a recently found class of consistent truncations of type IIB supergravity compactified on squashed five-dimensional Sasaki-Einstein manifolds. We derive the lower dimensional equations of motion and effective action, and comment on the supersymmetry of the resulting theory, which is consistent with N=4 gauged supergravity in d=5d=5, coupled to two vector multiplets. We compute fermion masses by linearizing around two AdS5AdS_{5} vacua of the theory: one that breaks N=4 down to N=2 spontaneously, and a second one which preserves no supersymmetries. The truncations under consideration are noteworthy in that they retain massive modes which are charged under a U(1) subgroup of the RR-symmetry, a feature that makes them interesting for applications to condensed matter phenomena via gauge/gravity duality. In this light, as an application of our general results we exhibit the coupling of the fermions to the type IIB holographic superconductor, and find a consistent further truncation of the fermion sector that retains a single spin-1/2 mode.Comment: 43 pages, 2 figures, PDFLaTeX; v2: added references, typos corrected, minor change
    corecore