1,619 research outputs found
Churn Intent Detection in Multilingual Chatbot Conversations and Social Media
We propose a new method to detect when users express the intent to leave a
service, also known as churn. While previous work focuses solely on social
media, we show that this intent can be detected in chatbot conversations. As
companies increasingly rely on chatbots they need an overview of potentially
churny users. To this end, we crowdsource and publish a dataset of churn intent
expressions in chatbot interactions in German and English. We show that
classifiers trained on social media data can detect the same intent in the
context of chatbots.
We introduce a classification architecture that outperforms existing work on
churn intent detection in social media. Moreover, we show that, using bilingual
word embeddings, a system trained on combined English and German data
outperforms monolingual approaches. As the only existing dataset is in English,
we crowdsource and publish a novel dataset of German tweets. We thus underline
the universal aspect of the problem, as examples of churn intent in English
help us identify churn in German tweets and chatbot conversations.Comment: 10 page
Coupled Representation Learning for Domains, Intents and Slots in Spoken Language Understanding
Representation learning is an essential problem in a wide range of
applications and it is important for performing downstream tasks successfully.
In this paper, we propose a new model that learns coupled representations of
domains, intents, and slots by taking advantage of their hierarchical
dependency in a Spoken Language Understanding system. Our proposed model learns
the vector representation of intents based on the slots tied to these intents
by aggregating the representations of the slots. Similarly, the vector
representation of a domain is learned by aggregating the representations of the
intents tied to a specific domain. To the best of our knowledge, it is the
first approach to jointly learning the representations of domains, intents, and
slots using their hierarchical relationships. The experimental results
demonstrate the effectiveness of the representations learned by our model, as
evidenced by improved performance on the contextual cross-domain reranking
task.Comment: IEEE SLT 201
Cross-lingual transfer learning for spoken language understanding
Typically, spoken language understanding (SLU) models are trained on
annotated data which are costly to gather. Aiming to reduce data needs for
bootstrapping a SLU system for a new language, we present a simple but
effective weight transfer approach using data from another language. The
approach is evaluated with our promising multi-task SLU framework developed
towards different languages. We evaluate our approach on the ATIS and a
real-world SLU dataset, showing that i) our monolingual models outperform the
state-of-the-art, ii) we can reduce data amounts needed for bootstrapping a SLU
system for a new language greatly, and iii) while multitask training improves
over separate training, different weight transfer settings may work best for
different SLU modules.Comment: accepted at ICASSP, 201
Embedding Grammars
Classic grammars and regular expressions can be used for a variety of
purposes, including parsing, intent detection, and matching. However, the
comparisons are performed at a structural level, with constituent elements
(words or characters) matched exactly. Recent advances in word embeddings show
that semantically related words share common features in a vector-space
representation, suggesting the possibility of a hybrid grammar and word
embedding. In this paper, we blend the structure of standard context-free
grammars with the semantic generalization capabilities of word embeddings to
create hybrid semantic grammars. These semantic grammars generalize the
specific terminals used by the programmer to other words and phrases with
related meanings, allowing the construction of compact grammars that match an
entire region of the vector space rather than matching specific elements
Outlier Detection for Improved Data Quality and Diversity in Dialog Systems
In a corpus of data, outliers are either errors: mistakes in the data that
are counterproductive, or are unique: informative samples that improve model
robustness. Identifying outliers can lead to better datasets by (1) removing
noise in datasets and (2) guiding collection of additional data to fill gaps.
However, the problem of detecting both outlier types has received relatively
little attention in NLP, particularly for dialog systems. We introduce a simple
and effective technique for detecting both erroneous and unique samples in a
corpus of short texts using neural sentence embeddings combined with
distance-based outlier detection. We also present a novel data collection
pipeline built atop our detection technique to automatically and iteratively
mine unique data samples while discarding erroneous samples. Experiments show
that our outlier detection technique is effective at finding errors while our
data collection pipeline yields highly diverse corpora that in turn produce
more robust intent classification and slot-filling models.Comment: Accepted as long paper to NAACL 201
Towards Open Intent Discovery for Conversational Text
Detecting and identifying user intent from text, both written and spoken,
plays an important role in modelling and understand dialogs. Existing research
for intent discovery model it as a classification task with a predefined set of
known categories. To generailze beyond these preexisting classes, we define a
new task of \textit{open intent discovery}. We investigate how intent can be
generalized to those not seen during training. To this end, we propose a
two-stage approach to this task - predicting whether an utterance contains an
intent, and then tagging the intent in the input utterance. Our model consists
of a bidirectional LSTM with a CRF on top to capture contextual semantics,
subject to some constraints. Self-attention is used to learn long distance
dependencies. Further, we adapt an adversarial training approach to improve
robustness and perforamce across domains. We also present a dataset of 25k
real-life utterances that have been labelled via crowd sourcing. Our
experiments across different domains and real-world datasets show the
effectiveness of our approach, with less than 100 annotated examples needed per
unique domain to recognize diverse intents. The approach outperforms
state-of-the-art baselines by 5-15% F1 score points
Learning Word Relatedness over Time
Search systems are often focused on providing relevant results for the "now",
assuming both corpora and user needs that focus on the present. However, many
corpora today reflect significant longitudinal collections ranging from 20
years of the Web to hundreds of years of digitized newspapers and books.
Understanding the temporal intent of the user and retrieving the most relevant
historical content has become a significant challenge. Common search features,
such as query expansion, leverage the relationship between terms but cannot
function well across all times when relationships vary temporally. In this
work, we introduce a temporal relationship model that is extracted from
longitudinal data collections. The model supports the task of identifying,
given two words, when they relate to each other. We present an algorithmic
framework for this task and show its application for the task of query
expansion, achieving high gain.Comment: 11 pages, EMNLP 201
We Built a Fake News & Click-bait Filter: What Happened Next Will Blow Your Mind!
It is completely amazing! Fake news and click-baits have totally invaded the
cyber space. Let us face it: everybody hates them for three simple reasons.
Reason #2 will absolutely amaze you. What these can achieve at the time of
election will completely blow your mind! Now, we all agree, this cannot go on,
you know, somebody has to stop it. So, we did this research on fake
news/click-bait detection and trust us, it is totally great research, it really
is! Make no mistake. This is the best research ever! Seriously, come have a
look, we have it all: neural networks, attention mechanism, sentiment lexicons,
author profiling, you name it. Lexical features, semantic features, we
absolutely have it all. And we have totally tested it, trust us! We have
results, and numbers, really big numbers. The best numbers ever! Oh, and
analysis, absolutely top notch analysis. Interested? Come read the shocking
truth about fake news and click-bait in the Bulgarian cyber space. You won't
believe what we have found!Comment: RANLP'2017, 7 pages, 1 figur
Parsing Coordination for Spoken Language Understanding
Typical spoken language understanding systems provide narrow semantic parses
using a domain-specific ontology. The parses contain intents and slots that are
directly consumed by downstream domain applications. In this work we discuss
expanding such systems to handle compound entities and intents by introducing a
domain-agnostic shallow parser that handles linguistic coordination. We show
that our model for parsing coordination learns domain-independent and
slot-independent features and is able to segment conjunct boundaries of many
different phrasal categories. We also show that using adversarial training can
be effective for improving generalization across different slot types for
coordination parsing.Comment: The paper was published in SLT 2018 conferenc
Subword Semantic Hashing for Intent Classification on Small Datasets
In this paper, we introduce the use of Semantic Hashing as embedding for the
task of Intent Classification and achieve state-of-the-art performance on three
frequently used benchmarks. Intent Classification on a small dataset is a
challenging task for data-hungry state-of-the-art Deep Learning based systems.
Semantic Hashing is an attempt to overcome such a challenge and learn robust
text classification. Current word embedding based are dependent on
vocabularies. One of the major drawbacks of such methods is out-of-vocabulary
terms, especially when having small training datasets and using a wider
vocabulary. This is the case in Intent Classification for chatbots, where
typically small datasets are extracted from internet communication. Two
problems arise by the use of internet communication. First, such datasets miss
a lot of terms in the vocabulary to use word embeddings efficiently. Second,
users frequently make spelling errors. Typically, the models for intent
classification are not trained with spelling errors and it is difficult to
think about ways in which users will make mistakes. Models depending on a word
vocabulary will always face such issues. An ideal classifier should handle
spelling errors inherently. With Semantic Hashing, we overcome these challenges
and achieve state-of-the-art results on three datasets: AskUbuntu, Chatbot, and
Web Application. Our benchmarks are available online:
https://github.com/kumar-shridhar/Know-Your-IntentComment: Accepted at IJCNN 2019 (Oral Presentation
- …