1,133 research outputs found

    Fisher Vectors Derived from Hybrid Gaussian-Laplacian Mixture Models for Image Annotation

    Full text link
    In the traditional object recognition pipeline, descriptors are densely sampled over an image, pooled into a high dimensional non-linear representation and then passed to a classifier. In recent years, Fisher Vectors have proven empirically to be the leading representation for a large variety of applications. The Fisher Vector is typically taken as the gradients of the log-likelihood of descriptors, with respect to the parameters of a Gaussian Mixture Model (GMM). Motivated by the assumption that different distributions should be applied for different datasets, we present two other Mixture Models and derive their Expectation-Maximization and Fisher Vector expressions. The first is a Laplacian Mixture Model (LMM), which is based on the Laplacian distribution. The second Mixture Model presented is a Hybrid Gaussian-Laplacian Mixture Model (HGLMM) which is based on a weighted geometric mean of the Gaussian and Laplacian distribution. An interesting property of the Expectation-Maximization algorithm for the latter is that in the maximization step, each dimension in each component is chosen to be either a Gaussian or a Laplacian. Finally, by using the new Fisher Vectors derived from HGLMMs, we achieve state-of-the-art results for both the image annotation and the image search by a sentence tasks.Comment: new version includes text synthesis by an RNN and experiments with the COCO benchmar

    Joint Learning of Correlated Sequence Labelling Tasks Using Bidirectional Recurrent Neural Networks

    Full text link
    The stream of words produced by Automatic Speech Recognition (ASR) systems is typically devoid of punctuations and formatting. Most natural language processing applications expect segmented and well-formatted texts as input, which is not available in ASR output. This paper proposes a novel technique of jointly modeling multiple correlated tasks such as punctuation and capitalization using bidirectional recurrent neural networks, which leads to improved performance for each of these tasks. This method could be extended for joint modeling of any other correlated sequence labeling tasks.Comment: Accepted in Interspeech 201

    Stark-Many body localization in interacting infinite dimensional systems

    Full text link
    We study bulk particle transport in a Fermi-Hubbard model on an infinite-dimensional Bethe lattice, driven by a constant electric field. Previous numerical studies showed that one dimensional analogs of this system exhibit a breakdown of diffusion due to Stark many-body localization (Stark-MBL) at least up to time which scales exponentially with the system size. Here, we consider systems initially in a spin density wave state using a combination of numerically exact and approximate techniques. We show that for sufficiently weak electric fields, the wave's momentum component decays exponentially with time in a way consistent with normal diffusion. By studying different wavelengths, we extract the dynamical exponent and the generalized diffusion coefficient at each field strength. Interestingly, we find a non-monotonic dependence of the dynamical exponent on the electric field. As the field increases towards a critical value proportional to the Hubbard interaction strength, transport slows down, becoming sub-diffusive. At large interaction strengths, however, transport speeds up again with increasing field, exhibiting super-diffusive characteristics when the electric field is comparable to the interaction strength. Eventually, at the large field limit, localization occurs and the current through the system is suppressed
    • …
    corecore