1,106 research outputs found
From Explainability to Explanation: Using a Dialogue Setting to Elicit Annotations with Justifications
Attari N, Heckmann M, Schlangen D. From Explainability to Explanation: Using a Dialogue Setting to Elicit Annotations with Justifications. In: Ultes S, ed. 20th Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDIAL 2019). Proceedings of the Conference, 11-13 September 2019 Stockholm, Sweden. Stroudsburg, PA: Association for Computational Linguistics (ACL); 2019: 331-335
N-best Response-based Analysis of Contradiction-awareness in Neural Response Generation Models
Avoiding the generation of responses that contradict the preceding context is
a significant challenge in dialogue response generation. One feasible method is
post-processing, such as filtering out contradicting responses from a resulting
n-best response list. In this scenario, the quality of the n-best list
considerably affects the occurrence of contradictions because the final
response is chosen from this n-best list. This study quantitatively analyzes
the contextual contradiction-awareness of neural response generation models
using the consistency of the n-best lists. Particularly, we used polar
questions as stimulus inputs for concise and quantitative analyses. Our tests
illustrate the contradiction-awareness of recent neural response generation
models and methodologies, followed by a discussion of their properties and
limitations.Comment: 8 pages, Accepted to The 23rd Annual Meeting of the Special Interest
Group on Discourse and Dialogue (SIGDIAL 2022
Probabilistic Dialogue Models with Prior Domain Knowledge
Probabilistic models such as Bayesian Networks are now in widespread use in spoken dialogue systems, but their scalability to complex interaction domains remains a challenge. One central limitation is that the state space of such models grows exponentially with the problem size, which makes parameter estimation increasingly difficult, especially for domains where only limited training data is available. In this paper, we show how to capture the underlying structure of a dialogue domain in terms of probabilistic rules operating on the dialogue state. The probabilistic rules are associated with a small, compact set of parameters that can be directly estimated from data. We argue that the introduction of this abstraction mechanism yields probabilistic models that are easier to learn and generalise better than their unstructured counterparts. We empirically demonstrate the benefits of such an approach learning a dialogue policy for a human-robot interaction domain based on a Wizard-of-Oz data set.
Proceedings of the 13th Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDIAL), pages 179–188, Seoul, South Korea, 5-6 July 2012
Contextual Out-of-Domain Utterance Handling With Counterfeit Data Augmentation
Neural dialog models often lack robustness to anomalous user input and
produce inappropriate responses which leads to frustrating user experience.
Although there are a set of prior approaches to out-of-domain (OOD) utterance
detection, they share a few restrictions: they rely on OOD data or multiple
sub-domains, and their OOD detection is context-independent which leads to
suboptimal performance in a dialog. The goal of this paper is to propose a
novel OOD detection method that does not require OOD data by utilizing
counterfeit OOD turns in the context of a dialog. For the sake of fostering
further research, we also release new dialog datasets which are 3 publicly
available dialog corpora augmented with OOD turns in a controllable way. Our
method outperforms state-of-the-art dialog models equipped with a conventional
OOD detection mechanism by a large margin in the presence of OOD utterances.Comment: ICASSP 201
- …