2,771 research outputs found

    Efficient training algorithms for HMMs using incremental estimation

    Get PDF
    Typically, parameter estimation for a hidden Markov model (HMM) is performed using an expectation-maximization (EM) algorithm with the maximum-likelihood (ML) criterion. The EM algorithm is an iterative scheme that is well-defined and numerically stable, but convergence may require a large number of iterations. For speech recognition systems utilizing large amounts of training material, this results in long training times. This paper presents an incremental estimation approach to speed-up the training of HMMs without any loss of recognition performance. The algorithm selects a subset of data from the training set, updates the model parameters based on the subset, and then iterates the process until convergence of the parameters. The advantage of this approach is a substantial increase in the number of iterations of the EM algorithm per training token, which leads to faster training. In order to achieve reliable estimation from a small fraction of the complete data set at each iteration, two training criteria are studied; ML and maximum a posteriori (MAP) estimation. Experimental results show that the training of the incremental algorithms is substantially faster than the conventional (batch) method and suffers no loss of recognition performance. Furthermore, the incremental MAP based training algorithm improves performance over the batch versio

    False discovery rate: setting the probability of false claim of detection

    Full text link
    When testing multiple hypothesis in a survey --e.g. many different source locations, template waveforms, and so on-- the final result consists in a set of confidence intervals, each one at a desired confidence level. But the probability that at least one of these intervals does not cover the true value increases with the number of trials. With a sufficiently large array of confidence intervals, one can be sure that at least one is missing the true value. In particular, the probability of false claim of detection becomes not negligible. In order to compensate for this, one should increase the confidence level, at the price of a reduced detection power. False discovery rate control is a relatively new statistical procedure that bounds the number of mistakes made when performing multiple hypothesis tests. We shall review this method, discussing exercise applications to the field of gravitational wave surveys.Comment: 7 pages, 3 table, 3 figures. Prepared for the Proceedings of GWDAW 9 (http://lappc-in39.in2p3.fr/GWDAW9) A new section was added with a numerical example, along with two tables and a figure related to the new section. Many smaller revisions to improve readibilit

    Energy Density of Non-Minimally Coupled Scalar Field Cosmologies

    Get PDF
    Scalar fields coupled to gravity via ξRΦ2\xi R {\Phi}^2 in arbitrary Friedmann-Robertson-Walker backgrounds can be represented by an effective flat space field theory. We derive an expression for the scalar energy density where the effective scalar mass becomes an explicit function of ξ\xi and the scale factor. The scalar quartic self-coupling gets shifted and can vanish for a particular choice of ξ\xi. Gravitationally induced symmetry breaking and de-stabilization are possible in this theory.Comment: 18 pages in standard Late

    Heat kernel regularization of the effective action for stochastic reaction-diffusion equations

    Full text link
    The presence of fluctuations and non-linear interactions can lead to scale dependence in the parameters appearing in stochastic differential equations. Stochastic dynamics can be formulated in terms of functional integrals. In this paper we apply the heat kernel method to study the short distance renormalizability of a stochastic (polynomial) reaction-diffusion equation with real additive noise. We calculate the one-loop {\emph{effective action}} and its ultraviolet scale dependent divergences. We show that for white noise a polynomial reaction-diffusion equation is one-loop {\emph{finite}} in d=0d=0 and d=1d=1, and is one-loop renormalizable in d=2d=2 and d=3d=3 space dimensions. We obtain the one-loop renormalization group equations and find they run with scale only in d=2d=2.Comment: 21 pages, uses ReV-TeX 3.

    Size Gap for Zero Temperature Black Holes in Semiclassical Gravity

    Get PDF
    We show that a gap exists in the allowed sizes of all zero temperature static spherically symmetric black holes in semiclassical gravity when only conformally invariant fields are present. The result holds for both charged and uncharged black holes. By size we mean the proper area of the event horizon. The range of sizes that do not occur depends on the numbers and types of quantized fields that are present. We also derive some general properties that both zero and nonzero temperature black holes have in all classical and semiclassical metric theories of gravity.Comment: 4 pages, ReVTeX, no figure

    Vibrations and fractional vibrations of rods, plates and Fresnel pseudo-processes

    Full text link
    Different initial and boundary value problems for the equation of vibrations of rods (also called Fresnel equation) are solved by exploiting the connection with Brownian motion and the heat equation. The analysis of the fractional version (of order ν\nu) of the Fresnel equation is also performed and, in detail, some specific cases, like ν=1/2\nu=1/2, 1/3, 2/3, are analyzed. By means of the fundamental solution of the Fresnel equation, a pseudo-process F(t)F(t), t>0t>0 with real sign-varying density is constructed and some of its properties examined. The equation of vibrations of plates is considered and the case of circular vibrating disks CRC_R is investigated by applying the methods of planar orthogonally reflecting Brownian motion within CRC_R. The composition of F with reflecting Brownian motion BB yields the law of biquadratic heat equation while the composition of FF with the first passage time TtT_t of BB produces a genuine probability law strictly connected with the Cauchy process.Comment: 33 pages,8 figure

    Flavor Changing Processes in Supersymmetric Models with Hybrid Gauge- and Gravity-Mediation

    Full text link
    We consider supersymmetric models where gauge mediation provides the dominant contributions to the soft supersymmetry breaking terms while gravity mediation provides sub-dominant yet non-negligible contributions. We further assume that the gravity-mediated contributions are subject to selection rules that follow from a Froggatt-Nielsen symmetry. This class of models constitutes an example of viable and natural non-minimally flavor violating models. The constraints from mixing in the neutral K system imply that the modifications to the Standard Model predictions for mixing in the neutral B_d and B_s systems are generically at most at the percent level, but can be of order ten percent for large tanβ\tan \beta. The modifications for the neutral D system mixing are generically at most of order a few percent, but in a special subclass of models they can be of order one. We point out ΔB=1\Delta B=1 processes relevant for flavor violation in hybrid mediation.Comment: 30 pages, 3 figure

    New constraints on dark matter from superconducting nanowires

    Get PDF
    Superconducting nanowires, a mature technology originally developed for quantum sensing, can be used as a target and sensor with which to search for dark matter interactions with electrons. Here we report on a 180-hour measurement of a tungsten silicide superconducting nanowire device with a mass of 4.3 nanograms. We use this to place new constraints on dark matter--electron interactions, including the strongest terrestrial constraints to date on sub-MeV (sub-eV) dark matter that interacts with electrons via scattering (absorption) processes.Comment: 5 pages + references, 5 figures. Includes supplementary material (4 pages, no figures). Added reference

    From spacetime foam to holographic foam cosmology

    Get PDF
    Due to quantum fluctuations, spacetime is foamy on small scales. For maximum spatial resolution of the geometry of spacetime, the holographic model of spacetime foam stipulates that the uncertainty or fluctuation of distance ll is given, on the average, by (llP2)1/3(l l_P^2)^{1/3} where lPl_P is the Planck length. Applied to cosmology, it predicts that the cosmic energy is of critical density and the cosmic entropy is the maximum allowed by the holographic principle. In addition, it requires the existence of unconventional (dark) energy/matter and accelerating cosmic expansion in the present era. We will argue that a holographic foam cosmology of this type has the potential to become a full fledged competitor (with distinct testable consequences) for scalar driven inflation.Comment: 8 pages, TeX; dedicated to Rafael Sorki
    corecore