1,133 research outputs found
Fisher Vectors Derived from Hybrid Gaussian-Laplacian Mixture Models for Image Annotation
In the traditional object recognition pipeline, descriptors are densely
sampled over an image, pooled into a high dimensional non-linear representation
and then passed to a classifier. In recent years, Fisher Vectors have proven
empirically to be the leading representation for a large variety of
applications. The Fisher Vector is typically taken as the gradients of the
log-likelihood of descriptors, with respect to the parameters of a Gaussian
Mixture Model (GMM). Motivated by the assumption that different distributions
should be applied for different datasets, we present two other Mixture Models
and derive their Expectation-Maximization and Fisher Vector expressions. The
first is a Laplacian Mixture Model (LMM), which is based on the Laplacian
distribution. The second Mixture Model presented is a Hybrid Gaussian-Laplacian
Mixture Model (HGLMM) which is based on a weighted geometric mean of the
Gaussian and Laplacian distribution. An interesting property of the
Expectation-Maximization algorithm for the latter is that in the maximization
step, each dimension in each component is chosen to be either a Gaussian or a
Laplacian. Finally, by using the new Fisher Vectors derived from HGLMMs, we
achieve state-of-the-art results for both the image annotation and the image
search by a sentence tasks.Comment: new version includes text synthesis by an RNN and experiments with
the COCO benchmar
Joint Learning of Correlated Sequence Labelling Tasks Using Bidirectional Recurrent Neural Networks
The stream of words produced by Automatic Speech Recognition (ASR) systems is
typically devoid of punctuations and formatting. Most natural language
processing applications expect segmented and well-formatted texts as input,
which is not available in ASR output. This paper proposes a novel technique of
jointly modeling multiple correlated tasks such as punctuation and
capitalization using bidirectional recurrent neural networks, which leads to
improved performance for each of these tasks. This method could be extended for
joint modeling of any other correlated sequence labeling tasks.Comment: Accepted in Interspeech 201
Stark-Many body localization in interacting infinite dimensional systems
We study bulk particle transport in a Fermi-Hubbard model on an
infinite-dimensional Bethe lattice, driven by a constant electric field.
Previous numerical studies showed that one dimensional analogs of this system
exhibit a breakdown of diffusion due to Stark many-body localization
(Stark-MBL) at least up to time which scales exponentially with the system
size. Here, we consider systems initially in a spin density wave state using a
combination of numerically exact and approximate techniques. We show that for
sufficiently weak electric fields, the wave's momentum component decays
exponentially with time in a way consistent with normal diffusion. By studying
different wavelengths, we extract the dynamical exponent and the generalized
diffusion coefficient at each field strength. Interestingly, we find a
non-monotonic dependence of the dynamical exponent on the electric field. As
the field increases towards a critical value proportional to the Hubbard
interaction strength, transport slows down, becoming sub-diffusive. At large
interaction strengths, however, transport speeds up again with increasing
field, exhibiting super-diffusive characteristics when the electric field is
comparable to the interaction strength. Eventually, at the large field limit,
localization occurs and the current through the system is suppressed
- …