438 research outputs found
Advances in Joint CTC-Attention based End-to-End Speech Recognition with a Deep CNN Encoder and RNN-LM
We present a state-of-the-art end-to-end Automatic Speech Recognition (ASR)
model. We learn to listen and write characters with a joint Connectionist
Temporal Classification (CTC) and attention-based encoder-decoder network. The
encoder is a deep Convolutional Neural Network (CNN) based on the VGG network.
The CTC network sits on top of the encoder and is jointly trained with the
attention-based decoder. During the beam search process, we combine the CTC
predictions, the attention-based decoder predictions and a separately trained
LSTM language model. We achieve a 5-10\% error reduction compared to prior
systems on spontaneous Japanese and Chinese speech, and our end-to-end model
beats out traditional hybrid ASR systems.Comment: Accepted for INTERSPEECH 201
Applying Deep Bidirectional LSTM and Mixture Density Network for Basketball Trajectory Prediction
Data analytics helps basketball teams to create tactics. However, manual data
collection and analytics are costly and ineffective. Therefore, we applied a
deep bidirectional long short-term memory (BLSTM) and mixture density network
(MDN) approach. This model is not only capable of predicting a basketball
trajectory based on real data, but it also can generate new trajectory samples.
It is an excellent application to help coaches and players decide when and
where to shoot. Its structure is particularly suitable for dealing with time
series problems. BLSTM receives forward and backward information at the same
time, while stacking multiple BLSTMs further increases the learning ability of
the model. Combined with BLSTMs, MDN is used to generate a multi-modal
distribution of outputs. Thus, the proposed model can, in principle, represent
arbitrary conditional probability distributions of output variables. We tested
our model with two experiments on three-pointer datasets from NBA SportVu data.
In the hit-or-miss classification experiment, the proposed model outperformed
other models in terms of the convergence speed and accuracy. In the trajectory
generation experiment, eight model-generated trajectories at a given time
closely matched real trajectories
Learning Spatial-Semantic Context with Fully Convolutional Recurrent Network for Online Handwritten Chinese Text Recognition
Online handwritten Chinese text recognition (OHCTR) is a challenging problem
as it involves a large-scale character set, ambiguous segmentation, and
variable-length input sequences. In this paper, we exploit the outstanding
capability of path signature to translate online pen-tip trajectories into
informative signature feature maps using a sliding window-based method,
successfully capturing the analytic and geometric properties of pen strokes
with strong local invariance and robustness. A multi-spatial-context fully
convolutional recurrent network (MCFCRN) is proposed to exploit the multiple
spatial contexts from the signature feature maps and generate a prediction
sequence while completely avoiding the difficult segmentation problem.
Furthermore, an implicit language model is developed to make predictions based
on semantic context within a predicting feature sequence, providing a new
perspective for incorporating lexicon constraints and prior knowledge about a
certain language in the recognition procedure. Experiments on two standard
benchmarks, Dataset-CASIA and Dataset-ICDAR, yielded outstanding results, with
correct rates of 97.10% and 97.15%, respectively, which are significantly
better than the best result reported thus far in the literature.Comment: 14 pages, 9 figure
- …