97 research outputs found
Towards More Efficient Insertion Transformer with Fractional Positional Encoding
Auto-regressive neural sequence models have been shown to be effective across
text generation tasks. However, their left-to-right decoding order prevents
generation from being parallelized. Insertion Transformer (Stern et al., 2019)
is an attractive alternative that allows outputting multiple tokens in a single
generation step. Nevertheless, due to the incompatibility between absolute
positional encoding and insertion-based generation schemes, it needs to refresh
the encoding of every token in the generated partial hypothesis at each step,
which could be costly. We design a novel reusable positional encoding scheme
for insertion transformers called Fractional Positional Encoding (FPE), which
allows reusing representations calculated in previous steps. Empirical studies
on various text generation tasks demonstrate the effectiveness of FPE, which
leads to floating-point operation reduction and latency improvements on batched
decoding
A Neural Network Approach to Context-Sensitive Generation of Conversational Responses
We present a novel response generation system that can be trained end to end
on large quantities of unstructured Twitter conversations. A neural network
architecture is used to address sparsity issues that arise when integrating
contextual information into classic statistical models, allowing the system to
take into account previous dialog utterances. Our dynamic-context generative
models show consistent gains over both context-sensitive and
non-context-sensitive Machine Translation and Information Retrieval baselines.Comment: A. Sordoni, M. Galley, M. Auli, C. Brockett, Y. Ji, M. Mitchell,
J.-Y. Nie, J. Gao, B. Dolan. 2015. A Neural Network Approach to
Context-Sensitive Generation of Conversational Responses. In Proc. of
NAACL-HLT. Pages 196-20
- …