3 research outputs found
Naturalizing a Programming Language via Interactive Learning
Our goal is to create a convenient natural language interface for performing
well-specified but complex actions such as analyzing data, manipulating text,
and querying databases. However, existing natural language interfaces for such
tasks are quite primitive compared to the power one wields with a programming
language. To bridge this gap, we start with a core programming language and
allow users to "naturalize" the core language incrementally by defining
alternative, more natural syntax and increasingly complex concepts in terms of
compositions of simpler ones. In a voxel world, we show that a community of
users can simultaneously teach a common system a diverse language and use it to
build hundreds of complex voxel structures. Over the course of three days,
these users went from using only the core language to using the naturalized
language in 85.9\% of the last 10K utterances.Comment: 10 pages, ACL201
Learning Personalized User Preference from Cold Start in Multi-turn Conversations
This paper presents a novel teachable conversation interaction system that is
capable of learning users preferences from cold start by gradually adapting to
personal preferences. In particular, the TAI system is able to automatically
identify and label user preference in live interactions, manage dialogue flows
for interactive teaching sessions, and reuse learned preference for preference
elicitation. We develop the TAI system by leveraging BERT encoder models to
encode both dialogue and relevant context information, and build action
prediction (AP), argument filling (AF) and named entity recognition (NER)
models to understand the teaching session. We adopt a seeker-provider
interaction loop mechanism to generate diverse dialogues from cold-start. TAI
is capable of learning user preference, which achieves 0.9122 turn level
accuracy on out-of-sample dataset, and has been successfully adopted in
production.Comment: preference, personalization, cold-start, dialogue, LLM. embeddin
Offline and Online Satisfaction Prediction in Open-Domain Conversational Systems
Predicting user satisfaction in conversational systems has become critical,
as spoken conversational assistants operate in increasingly complex domains.
Online satisfaction prediction (i.e., predicting satisfaction of the user with
the system after each turn) could be used as a new proxy for implicit user
feedback, and offers promising opportunities to create more responsive and
effective conversational agents, which adapt to the user's engagement with the
agent. To accomplish this goal, we propose a conversational satisfaction
prediction model specifically designed for open-domain spoken conversational
agents, called ConvSAT. To operate robustly across domains, ConvSAT aggregates
multiple representations of the conversation, namely the conversation history,
utterance and response content, and system- and user-oriented behavioral
signals. We first calibrate ConvSAT performance against state of the art
methods on a standard dataset (Dialogue Breakdown Detection Challenge) in an
online regime, and then evaluate ConvSAT on a large dataset of conversations
with real users, collected as part of the Alexa Prize competition. Our
experimental results show that ConvSAT significantly improves satisfaction
prediction for both offline and online setting on both datasets, compared to
the previously reported state-of-the-art approaches. The insights from our
study can enable more intelligent conversational systems, which could adapt in
real-time to the inferred user satisfaction and engagement.Comment: Published in CIKM '19, 10 page