3,648 research outputs found
SF-DST: Few-Shot Self-Feeding Reading Comprehension Dialogue State Tracking with Auxiliary Task
Few-shot dialogue state tracking (DST) model tracks user requests in dialogue
with reliable accuracy even with a small amount of data. In this paper, we
introduce an ontology-free few-shot DST with self-feeding belief state input.
The self-feeding belief state input increases the accuracy in multi-turn
dialogue by summarizing previous dialogue. Also, we newly developed a slot-gate
auxiliary task. This new auxiliary task helps classify whether a slot is
mentioned in the dialogue. Our model achieved the best score in a few-shot
setting for four domains on multiWOZ 2.0.Comment: Accepted in INTERSPEECH 202
Attention-Informed Mixed-Language Training for Zero-shot Cross-lingual Task-oriented Dialogue Systems
Recently, data-driven task-oriented dialogue systems have achieved promising
performance in English. However, developing dialogue systems that support
low-resource languages remains a long-standing challenge due to the absence of
high-quality data. In order to circumvent the expensive and time-consuming data
collection, we introduce Attention-Informed Mixed-Language Training (MLT), a
novel zero-shot adaptation method for cross-lingual task-oriented dialogue
systems. It leverages very few task-related parallel word pairs to generate
code-switching sentences for learning the inter-lingual semantics across
languages. Instead of manually selecting the word pairs, we propose to extract
source words based on the scores computed by the attention layer of a trained
English task-related model and then generate word pairs using existing
bilingual dictionaries. Furthermore, intensive experiments with different
cross-lingual embeddings demonstrate the effectiveness of our approach.
Finally, with very few word pairs, our model achieves significant zero-shot
adaptation performance improvements in both cross-lingual dialogue state
tracking and natural language understanding (i.e., intent detection and slot
filling) tasks compared to the current state-of-the-art approaches, which
utilize a much larger amount of bilingual data.Comment: Accepted as an oral presentation in AAAI 202
- …