2,753 research outputs found
ASR error management for improving spoken language understanding
This paper addresses the problem of automatic speech recognition (ASR) error
detection and their use for improving spoken language understanding (SLU)
systems. In this study, the SLU task consists in automatically extracting, from
ASR transcriptions , semantic concepts and concept/values pairs in a e.g
touristic information system. An approach is proposed for enriching the set of
semantic labels with error specific labels and by using a recently proposed
neural approach based on word embeddings to compute well calibrated ASR
confidence measures. Experimental results are reported showing that it is
possible to decrease significantly the Concept/Value Error Rate with a state of
the art system, outperforming previously published results performance on the
same experimental data. It also shown that combining an SLU approach based on
conditional random fields with a neural encoder/decoder attention based
architecture , it is possible to effectively identifying confidence islands and
uncertain semantic output segments useful for deciding appropriate error
handling actions by the dialogue manager strategy .Comment: Interspeech 2017, Aug 2017, Stockholm, Sweden. 201
Patent Citation Dynamics Modeling via Multi-Attention Recurrent Networks
Modeling and forecasting forward citations to a patent is a central task for
the discovery of emerging technologies and for measuring the pulse of inventive
progress. Conventional methods for forecasting these forward citations cast the
problem as analysis of temporal point processes which rely on the conditional
intensity of previously received citations. Recent approaches model the
conditional intensity as a chain of recurrent neural networks to capture memory
dependency in hopes of reducing the restrictions of the parametric form of the
intensity function. For the problem of patent citations, we observe that
forecasting a patent's chain of citations benefits from not only the patent's
history itself but also from the historical citations of assignees and
inventors associated with that patent. In this paper, we propose a
sequence-to-sequence model which employs an attention-of-attention mechanism to
capture the dependencies of these multiple time sequences. Furthermore, the
proposed model is able to forecast both the timestamp and the category of a
patent's next citation. Extensive experiments on a large patent citation
dataset collected from USPTO demonstrate that the proposed model outperforms
state-of-the-art models at forward citation forecasting
Efficient Modeling of Future Context for Image Captioning
Existing approaches to image captioning usually generate the sentence
word-by-word from left to right, with the constraint of conditioned on local
context including the given image and history generated words. There have been
many studies target to make use of global information during decoding, e.g.,
iterative refinement. However, it is still under-explored how to effectively
and efficiently incorporate the future context. To respond to this issue,
inspired by that Non-Autoregressive Image Captioning (NAIC) can leverage
two-side relation with modified mask operation, we aim to graft this advance to
the conventional Autoregressive Image Captioning (AIC) model while maintaining
the inference efficiency without extra time cost. Specifically, AIC and NAIC
models are first trained combined with shared visual encoders, forcing the
visual encoder to contain sufficient and valid future context; then the AIC
model is encouraged to capture the causal dynamics of cross-layer interchanging
from NAIC model on its unconfident words, which follows a teacher-student
paradigm and optimized with the distribution calibration training objective.
Empirical evidences demonstrate that our proposed approach clearly surpass the
state-of-the-art baselines in both automatic metrics and human evaluations on
the MS COCO benchmark. The source code is available at:
https://github.com/feizc/Future-Caption.Comment: ACM Multimedia 202
- …