82 research outputs found
Filling Conversation Ellipsis for Better Social Dialog Understanding
The phenomenon of ellipsis is prevalent in social conversations. Ellipsis
increases the difficulty of a series of downstream language understanding
tasks, such as dialog act prediction and semantic role labeling. We propose to
resolve ellipsis through automatic sentence completion to improve language
understanding. However, automatic ellipsis completion can result in output
which does not accurately reflect user intent. To address this issue, we
propose a method which considers both the original utterance that has ellipsis
and the automatically completed utterance in dialog act and semantic role
labeling tasks. Specifically, we first complete user utterances to resolve
ellipsis using an end-to-end pointer network model. We then train a prediction
model using both utterances containing ellipsis and our automatically completed
utterances. Finally, we combine the prediction results from these two
utterances using a selection model that is guided by expert knowledge. Our
approach improves dialog act prediction and semantic role labeling by 1.3% and
2.5% in F1 score respectively in social conversations. We also present an
open-domain human-machine conversation dataset with manually completed user
utterances and annotated semantic role labeling after manual completion.Comment: Accepted to AAAI 202
Translate the Beauty in Songs: Jointly Learning to Align Melody and Translate Lyrics
Song translation requires both translation of lyrics and alignment of music
notes so that the resulting verse can be sung to the accompanying melody, which
is a challenging problem that has attracted some interests in different aspects
of the translation process. In this paper, we propose Lyrics-Melody Translation
with Adaptive Grouping (LTAG), a holistic solution to automatic song
translation by jointly modeling lyrics translation and lyrics-melody alignment.
It is a novel encoder-decoder framework that can simultaneously translate the
source lyrics and determine the number of aligned notes at each decoding step
through an adaptive note grouping module. To address data scarcity, we
commissioned a small amount of training data annotated specifically for this
task and used large amounts of augmented data through back-translation.
Experiments conducted on an English-Chinese song translation data set show the
effectiveness of our model in both automatic and human evaluation.Comment: 13 page
Exploiting sparseness in de novo genome assembly
Background: The very large memory requirements for the construction of assembly graphs for de novo genome assembly limit current algorithms to super-computing environments. Methods: In this paper, we demonstrate that constructing a sparse assembly graph which stores only a small fraction of the observed k- mers as nodes and the links between these nodes allows the de novo assembly of even moderately-sized genomes (~500 M) on a typical laptop computer. Results: We implement this sparse graph concept in a proof-of-principle software package, SparseAssembler, utilizing a new sparse k- mer graph structure evolved from the de Bruijn graph. We test our SparseAssembler with both simulated and real data, achieving ~90% memory savings and retaining high assembly accuracy, without sacrificing speed in comparison to existing de novo assemblers
DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation
We introduce DS-1000, a code generation benchmark with a thousand data
science problems spanning seven Python libraries, such as NumPy and Pandas.
Compared to prior works, DS-1000 incorporates three core features. First, our
problems reflect diverse, realistic, and practical use cases since we collected
them from StackOverflow. Second, our automatic evaluation is highly specific
(reliable) -- across all Codex-002-predicted solutions that our evaluation
accept, only 1.8% of them are incorrect; we achieve this with multi-criteria
metrics, checking both functional correctness by running test cases and
surface-form constraints by restricting API usages or keywords. Finally, we
proactively defend against memorization by slightly modifying our problems to
be different from the original StackOverflow source; consequently, models
cannot answer them correctly by memorizing the solutions from pre-training. The
current best public system (Codex-002) achieves 43.3% accuracy, leaving ample
room for improvement. We release our benchmark at
https://ds1000-code-gen.github.io
- …