1,338 research outputs found
Learning Personalized End-to-End Goal-Oriented Dialog
Most existing works on dialog systems only consider conversation content
while neglecting the personality of the user the bot is interacting with, which
begets several unsolved issues. In this paper, we present a personalized
end-to-end model in an attempt to leverage personalization in goal-oriented
dialogs. We first introduce a Profile Model which encodes user profiles into
distributed embeddings and refers to conversation history from other similar
users. Then a Preference Model captures user preferences over knowledge base
entities to handle the ambiguity in user requests. The two models are combined
into the Personalized MemN2N. Experiments show that the proposed model achieves
qualitative performance improvements over state-of-the-art methods. As for
human evaluation, it also outperforms other approaches in terms of task
completion rate and user satisfaction.Comment: Accepted by AAAI 201
Towards Coherent and Engaging Spoken Dialog Response Generation Using Automatic Conversation Evaluators
Encoder-decoder based neural architectures serve as the basis of
state-of-the-art approaches in end-to-end open domain dialog systems. Since
most of such systems are trained with a maximum likelihood~(MLE) objective they
suffer from issues such as lack of generalizability and the generic response
problem, i.e., a system response that can be an answer to a large number of
user utterances, e.g., "Maybe, I don't know." Having explicit feedback on the
relevance and interestingness of a system response at each turn can be a useful
signal for mitigating such issues and improving system quality by selecting
responses from different approaches. Towards this goal, we present a system
that evaluates chatbot responses at each dialog turn for coherence and
engagement. Our system provides explicit turn-level dialog quality feedback,
which we show to be highly correlated with human evaluation. To show that
incorporating this feedback in the neural response generation models improves
dialog quality, we present two different and complementary mechanisms to
incorporate explicit feedback into a neural response generation model:
reranking and direct modification of the loss function during training. Our
studies show that a response generation model that incorporates these combined
feedback mechanisms produce more engaging and coherent responses in an
open-domain spoken dialog setting, significantly improving the response quality
using both automatic and human evaluation
Learning from Dialogue after Deployment: Feed Yourself, Chatbot!
The majority of conversations a dialogue agent sees over its lifetime occur
after it has already been trained and deployed, leaving a vast store of
potential training signal untapped. In this work, we propose the self-feeding
chatbot, a dialogue agent with the ability to extract new training examples
from the conversations it participates in. As our agent engages in
conversation, it also estimates user satisfaction in its responses. When the
conversation appears to be going well, the user's responses become new training
examples to imitate. When the agent believes it has made a mistake, it asks for
feedback; learning to predict the feedback that will be given improves the
chatbot's dialogue abilities further. On the PersonaChat chit-chat dataset with
over 131k training examples, we find that learning from dialogue with a
self-feeding chatbot significantly improves performance, regardless of the
amount of traditional supervision.Comment: ACL 201
The Dialogue Dodecathlon: Open-Domain Knowledge and Image Grounded Conversational Agents
We introduce dodecaDialogue: a set of 12 tasks that measures if a
conversational agent can communicate engagingly with personality and empathy,
ask questions, answer questions by utilizing knowledge resources, discuss
topics and situations, and perceive and converse about images. By multi-tasking
on such a broad large-scale set of data, we hope to both move towards and
measure progress in producing a single unified agent that can perceive, reason
and converse with humans in an open-domain setting. We show that such
multi-tasking improves over a BERT pre-trained baseline, largely due to
multi-tasking with very large dialogue datasets in a similar domain, and that
the multi-tasking in general provides gains to both text and image-based tasks
using several metrics in both the fine-tune and task transfer settings. We
obtain state-of-the-art results on many of the tasks, providing a strong
baseline for this challenge.Comment: ACL 202
Designing for Health Chatbots
Building conversational agents have many technical, design and linguistic
challenges. Other more complex elements include using emotionally intelligent
conversational agent to build trust with the individuals. In this chapter, we
introduce the nature of conversational user interfaces (CUIs) for health and
describe UX design principles informed by a systematic literature review of
relevant research works. We analyze scientific literature in conversational
interfaces and chatterbots, providing a survey of major studies and describing
UX design principles and interaction patterns
User Intent Prediction in Information-seeking Conversations
Conversational assistants are being progressively adopted by the general
population. However, they are not capable of handling complicated
information-seeking tasks that involve multiple turns of information exchange.
Due to the limited communication bandwidth in conversational search, it is
important for conversational assistants to accurately detect and predict user
intent in information-seeking conversations. In this paper, we investigate two
aspects of user intent prediction in an information-seeking setting. First, we
extract features based on the content, structural, and sentiment
characteristics of a given utterance, and use classic machine learning methods
to perform user intent prediction. We then conduct an in-depth feature
importance analysis to identify key features in this prediction task. We find
that structural features contribute most to the prediction performance. Given
this finding, we construct neural classifiers to incorporate context
information and achieve better performance without feature engineering. Our
findings can provide insights into the important factors and effective methods
of user intent prediction in information-seeking conversations.Comment: Accepted to CHIIR 201
Tartan: A retrieval-based socialbot powered by a dynamic finite-state machine architecture
This paper describes the Tartan conversational agent built for the 2018 Alexa
Prize Competition. Tartan is a non-goal-oriented socialbot focused around
providing users with an engaging and fluent casual conversation. Tartan's key
features include an emphasis on structured conversation based on flexible
finite-state models and an approach focused on understanding and using
conversational acts. To provide engaging conversations, Tartan blends
script-like yet dynamic responses with data-based generative and retrieval
models. Unique to Tartan is that our dialog manager is modeled as a dynamic
Finite State Machine. To our knowledge, no other conversational agent
implementation has followed this specific structure
Dialogue Agents 101: A Beginner's Guide to Critical Ingredients for Designing Effective Conversational Systems
Sharing ideas through communication with peers is the primary mode of human
interaction. Consequently, extensive research has been conducted in the area of
conversational AI, leading to an increase in the availability and diversity of
conversational tasks, datasets, and methods. However, with numerous tasks being
explored simultaneously, the current landscape of conversational AI becomes
fragmented. Therefore, initiating a well-thought-out model for a dialogue agent
can pose significant challenges for a practitioner. Towards highlighting the
critical ingredients needed for a practitioner to design a dialogue agent from
scratch, the current study provides a comprehensive overview of the primary
characteristics of a dialogue agent, the supporting tasks, their corresponding
open-domain datasets, and the methods used to benchmark these datasets. We
observe that different methods have been used to tackle distinct dialogue
tasks. However, building separate models for each task is costly and does not
leverage the correlation among the several tasks of a dialogue agent. As a
result, recent trends suggest a shift towards building unified foundation
models. To this end, we propose UNIT, a UNified dIalogue dataseT constructed
from conversations of existing datasets for different dialogue tasks capturing
the nuances for each of them. We also examine the evaluation strategies used to
measure the performance of dialogue agents and highlight the scope for future
research in the area of conversational AI.Comment: 21 pages, 3 figures, 3 table
Multipurpose Intelligent Process Automation via Conversational Assistant
Intelligent Process Automation (IPA) is an emerging technology with a primary
goal to assist the knowledge worker by taking care of repetitive, routine and
low-cognitive tasks. Conversational agents that can interact with users in a
natural language are potential application for IPA systems. Such intelligent
agents can assist the user by answering specific questions and executing
routine tasks that are ordinarily performed in a natural language (i.e.,
customer support). In this work, we tackle a challenge of implementing an IPA
conversational assistant in a real-world industrial setting with a lack of
structured training data. Our proposed system brings two significant benefits:
First, it reduces repetitive and time-consuming activities and, therefore,
allows workers to focus on more intelligent processes. Second, by interacting
with users, it augments the resources with structured and to some extent
labeled training data. We showcase the usage of the latter by re-implementing
several components of our system with Transfer Learning (TL) methods.Comment: Presented at the AAAI-20 Workshop on Intelligent Process Automatio
- …