462 research outputs found
Deep Reinforcement Learning for Dialogue Generation
Recent neural models of dialogue generation offer great promise for
generating responses for conversational agents, but tend to be shortsighted,
predicting utterances one at a time while ignoring their influence on future
outcomes. Modeling the future direction of a dialogue is crucial to generating
coherent, interesting dialogues, a need which led traditional NLP models of
dialogue to draw on reinforcement learning. In this paper, we show how to
integrate these goals, applying deep reinforcement learning to model future
reward in chatbot dialogue. The model simulates dialogues between two virtual
agents, using policy gradient methods to reward sequences that display three
useful conversational properties: informativity (non-repetitive turns),
coherence, and ease of answering (related to forward-looking function). We
evaluate our model on diversity, length as well as with human judges, showing
that the proposed algorithm generates more interactive responses and manages to
foster a more sustained conversation in dialogue simulation. This work marks a
first step towards learning a neural conversational model based on the
long-term success of dialogues
MIRIAM: A Multimodal Chat-Based Interface for Autonomous Systems
We present MIRIAM (Multimodal Intelligent inteRactIon for Autonomous
systeMs), a multimodal interface to support situation awareness of autonomous
vehicles through chat-based interaction. The user is able to chat about the
vehicle's plan, objectives, previous activities and mission progress. The
system is mixed initiative in that it pro-actively sends messages about key
events, such as fault warnings. We will demonstrate MIRIAM using SeeByte's
SeeTrack command and control interface and Neptune autonomy simulator.Comment: 2 pages, ICMI'17, 19th ACM International Conference on Multimodal
Interaction, November 13-17 2017, Glasgow, U
A Review of Evaluation Techniques for Social Dialogue Systems
In contrast with goal-oriented dialogue, social dialogue has no clear measure
of task success. Consequently, evaluation of these systems is notoriously hard.
In this paper, we review current evaluation methods, focusing on automatic
metrics. We conclude that turn-based metrics often ignore the context and do
not account for the fact that several replies are valid, while end-of-dialogue
rewards are mainly hand-crafted. Both lack grounding in human perceptions.Comment: 2 page
An Ensemble Model with Ranking for Social Dialogue
Open-domain social dialogue is one of the long-standing goals of Artificial
Intelligence. This year, the Amazon Alexa Prize challenge was announced for the
first time, where real customers get to rate systems developed by leading
universities worldwide. The aim of the challenge is to converse "coherently and
engagingly with humans on popular topics for 20 minutes". We describe our Alexa
Prize system (called 'Alana') consisting of an ensemble of bots, combining
rule-based and machine learning systems, and using a contextual ranking
mechanism to choose a system response. The ranker was trained on real user
feedback received during the competition, where we address the problem of how
to train on the noisy and sparse feedback obtained during the competition.Comment: NIPS 2017 Workshop on Conversational A
- …