3,063 research outputs found
Unsupervised Online Multitask Learning of Behavioral Sentence Embeddings
Unsupervised learning has been an attractive method for easily deriving
meaningful data representations from vast amounts of unlabeled data. These
representations, or embeddings, often yield superior results in many tasks,
whether used directly or as features in subsequent training stages. However,
the quality of the embeddings is highly dependent on the assumed knowledge in
the unlabeled data and how the system extracts information without supervision.
Domain portability is also very limited in unsupervised learning, often
requiring re-training on other in-domain corpora to achieve robustness. In this
work we present a multitask paradigm for unsupervised contextual learning of
behavioral interactions which addresses unsupervised domain adaption. We
introduce an online multitask objective into unsupervised learning and show
that sentence embeddings generated through this process increases performance
of affective tasks
Facebook Reaction-Based Emotion Classifier as Cue for Sarcasm Detection
Online social media users react to content in them based on context. Emotions
or mood play a significant part of these reactions, which has filled these
platforms with opinionated content. Different approaches and applications to
make better use of this data are continuously being developed. However, due to
the nature of the data, the variety of platforms, and dynamic online user
behavior, there are still many issues to be dealt with. It remains a challenge
to properly obtain a reliable emotional status from a user prior to posting a
comment. This work introduces a methodology that explores semi-supervised
multilingual emotion detection based on the overlap of Facebook reactions and
textual data. With the resulting emotion detection system we evaluate the
possibility of using emotions and user behavior features for the task of
sarcasm detection. More than 1 million English and Chinese comments from over
62,000 public Facebook pages posts have been collected and processed, conducted
experiments show acceptable performance metrics.Comment: 10 pages ACM forma
Leveraging Semantic Web Search and Browse Sessions for Multi-Turn Spoken Dialog Systems
Training statistical dialog models in spoken dialog systems (SDS) requires
large amounts of annotated data. The lack of scalable methods for data mining
and annotation poses a significant hurdle for state-of-the-art statistical
dialog managers. This paper presents an approach that directly leverage
billions of web search and browse sessions to overcome this hurdle. The key
insight is that task completion through web search and browse sessions is (a)
predictable and (b) generalizes to spoken dialog task completion. The new
method automatically mines behavioral search and browse patterns from web logs
and translates them into spoken dialog models. We experiment with naturally
occurring spoken dialogs and large scale web logs. Our session-based models
outperform the state-of-the-art method for entity extraction task in SDS. We
also achieve better performance for both entity and relation extraction on web
search queries when compared with nontrivial baselines.Comment: ICASSP 201
Verisimilar Image Synthesis for Accurate Detection and Recognition of Texts in Scenes
The requirement of large amounts of annotated images has become one grand
challenge while training deep neural network models for various visual
detection and recognition tasks. This paper presents a novel image synthesis
technique that aims to generate a large amount of annotated scene text images
for training accurate and robust scene text detection and recognition models.
The proposed technique consists of three innovative designs. First, it realizes
"semantic coherent" synthesis by embedding texts at semantically sensible
regions within the background image, where the semantic coherence is achieved
by leveraging the semantic annotations of objects and image regions that have
been created in the prior semantic segmentation research. Second, it exploits
visual saliency to determine the embedding locations within each semantic
sensible region, which coincides with the fact that texts are often placed
around homogeneous regions for better visibility in scenes. Third, it designs
an adaptive text appearance model that determines the color and brightness of
embedded texts by learning from the feature of real scene text images
adaptively. The proposed technique has been evaluated over five public datasets
and the experiments show its superior performance in training accurate and
robust scene text detection and recognition models.Comment: 14 pages, ECCV2018, datasets:
https://github.com/fnzhan/Verisimilar-Image-Synthesis-for-Accurate-Detection-and-Recognition-of-Texts-in-Scene
Constructionist Steps Towards an Autonomously Empathetic System
Prior efforts to create an autonomous computer system capable of predicting
what a human being is thinking or feeling from facial expression data have been
largely based on outdated, inaccurate models of how emotions work that rely on
many scientifically questionable assumptions. In our research, we are creating
an empathetic system that incorporates the latest provable scientific
understanding of emotions: that they are constructs of the human mind, rather
than universal expressions of distinct internal states. Thus, our system uses a
user-dependent method of analysis and relies heavily on contextual information
to make predictions about what subjects are experiencing. Our system's accuracy
and therefore usefulness are built on provable ground truths that prohibit the
drawing of inaccurate conclusions that other systems could too easily make.Comment: Submitted for SIGCHI ICMI 2018's Late-Breaking-Work trac
Beneath the Tip of the Iceberg: Current Challenges and New Directions in Sentiment Analysis Research
Sentiment analysis as a field has come a long way since it was first
introduced as a task nearly 20 years ago. It has widespread commercial
applications in various domains like marketing, risk management, market
research, and politics, to name a few. Given its saturation in specific
subtasks -- such as sentiment polarity classification -- and datasets, there is
an underlying perception that this field has reached its maturity. In this
article, we discuss this perception by pointing out the shortcomings and
under-explored, yet key aspects of this field that are necessary to attain true
sentiment understanding. We analyze the significant leaps responsible for its
current relevance. Further, we attempt to chart a possible course for this
field that covers many overlooked and unanswered questions.Comment: Published in the IEEE Transactions on Affective Computing (TAFFC
Linking emotions to behaviors through deep transfer learning
Human behavior refers to the way humans act and interact. Understanding human
behavior is a cornerstone of observational practice, especially in
psychotherapy. An important cue of behavior analysis is the dynamical changes
of emotions during the conversation. Domain experts integrate emotional
information in a highly nonlinear manner, thus, it is challenging to explicitly
quantify the relationship between emotions and behaviors. In this work, we
employ deep transfer learning to analyze their inferential capacity and
contextual importance. We first train a network to quantify emotions from
acoustic signals and then use information from the emotion recognition network
as features for behavior recognition. We treat this emotion-related information
as behavioral primitives and further train higher level layers towards behavior
quantification. Through our analysis, we find that emotion-related information
is an important cue for behavior recognition. Further, we investigate the
importance of emotional-context in the expression of behavior by constraining
(or not) the neural networks' contextual view of the data. This demonstrates
that the sequence of emotions is critical in behavior expression. To achieve
these frameworks we employ hybrid architectures of convolutional networks and
recurrent networks to extract emotion-related behavior primitives and
facilitate automatic behavior recognition from speech.Comment: 23 pages, 8 figure
CASCADE: Contextual Sarcasm Detection in Online Discussion Forums
The literature in automated sarcasm detection has mainly focused on lexical,
syntactic and semantic-level analysis of text. However, a sarcastic sentence
can be expressed with contextual presumptions, background and commonsense
knowledge. In this paper, we propose CASCADE (a ContextuAl SarCasm DEtector)
that adopts a hybrid approach of both content and context-driven modeling for
sarcasm detection in online social media discussions. For the latter, CASCADE
aims at extracting contextual information from the discourse of a discussion
thread. Also, since the sarcastic nature and form of expression can vary from
person to person, CASCADE utilizes user embeddings that encode stylometric and
personality features of the users. When used along with content-based feature
extractors such as Convolutional Neural Networks (CNNs), we see a significant
boost in the classification performance on a large Reddit corpus.Comment: Accepted in COLING 201
Is it Safe to Drive? An Overview of Factors, Challenges, and Datasets for Driveability Assessment in Autonomous Driving
With recent advances in learning algorithms and hardware development,
autonomous cars have shown promise when operating in structured environments
under good driving conditions. However, for complex, cluttered and unseen
environments with high uncertainty, autonomous driving systems still frequently
demonstrate erroneous or unexpected behaviors, that could lead to catastrophic
outcomes. Autonomous vehicles should ideally adapt to driving conditions; while
this can be achieved through multiple routes, it would be beneficial as a first
step to be able to characterize Driveability in some quantified form. To this
end, this paper aims to create a framework for investigating different factors
that can impact driveability. Also, one of the main mechanisms to adapt
autonomous driving systems to any driving condition is to be able to learn and
generalize from representative scenarios. The machine learning algorithms that
currently do so learn predominantly in a supervised manner and consequently
need sufficient data for robust and efficient learning. Therefore, we also
perform a comparative overview of 45 public driving datasets that enable
learning and publish this dataset index at
https://sites.google.com/view/driveability-survey-datasets. Specifically, we
categorize the datasets according to use cases, and highlight the datasets that
capture complicated and hazardous driving conditions which can be better used
for training robust driving models. Furthermore, by discussions of what driving
scenarios are not covered by existing public datasets and what driveability
factors need more investigation and data acquisition, this paper aims to
encourage both targeted dataset collection and the proposal of novel
driveability metrics that enhance the robustness of autonomous cars in adverse
environments
ACII 2009: Affective Computing and Intelligent Interaction. Proceedings of the Doctoral Consortium
This volume collects the contributions presented at the ACII 2009 Doctoral Consortium, the event aimed at gathering PhD students with the goal of sharing ideas about the theories behind affective computing; its development; and its application. Published papers have been selected out a large number of high quality submissions covering a wide spectrum of topics including the analysis of human-human, human-machine and human-robot interactions, the analysis of physiology and nonverbal behavior in affective phenomena, the effect of emotions on language and spoken interaction, and the embodiment of affective behaviors
- …