99 research outputs found
Bayesian Inference of Social Norms as Shared Constraints on Behavior
People act upon their desires, but often, also act in adherence to implicit
social norms. How do people infer these unstated social norms from others'
behavior, especially in novel social contexts? We propose that laypeople have
intuitive theories of social norms as behavioral constraints shared across
different agents in the same social context. We formalize inference of norms
using a Bayesian Theory of Mind approach, and show that this computational
approach provides excellent predictions of how people infer norms in two
scenarios. Our results suggest that people separate the influence of norms and
individual desires on others' actions, and have implications for modelling
generalizations of hidden causes of behavior.Comment: 7 pages, 5 figures, to appear in CogSci 2019, code available at
https://github.com/ztangent/norms-cogsci1
Context-Guided BERT for Targeted Aspect-Based Sentiment Analysis
Aspect-based sentiment analysis (ABSA) and Targeted ASBA (TABSA) allow
finer-grained inferences about sentiment to be drawn from the same text,
depending on context. For example, a given text can have different targets
(e.g., neighborhoods) and different aspects (e.g., price or safety), with
different sentiment associated with each target-aspect pair. In this paper, we
investigate whether adding context to self-attention models improves
performance on (T)ABSA. We propose two variants of Context-Guided BERT
(CG-BERT) that learn to distribute attention under different contexts. We first
adapt a context-aware Transformer to produce a CG-BERT that uses context-guided
softmax-attention. Next, we propose an improved Quasi-Attention CG-BERT model
that learns a compositional attention that supports subtractive attention. We
train both models with pretrained BERT on two (T)ABSA datasets: SentiHood and
SemEval-2014 (Task 4). Both models achieve new state-of-the-art results with
our QACG-BERT model having the best performance. Furthermore, we provide
analyses of the impact of context in the our proposed models. Our work provides
more evidence for the utility of adding context-dependencies to pretrained
self-attention-based language models for context-based natural language tasks
Recommended from our members
Pragmatically Informative Color Generation by Grounding Contextual Modifiers
Grounding language in contextual information is crucial for fine-grained natural language understanding. One important task that involves grounding contextual modifiers is color generation. Given a reference color green, and a modifier bluey, how does one generate a color that could represent bluey green? We propose a computational pragmatics model that formulates this color generation task as a recursive game between speakers and listeners. In our model, a pragmatic speaker reasons about the inferences that a listener would make, and thus generates a modified color that is maximally informative to help the listener recover the original referents. In this paper, we show that incorporating pragmatic information provides significant improvements in performance compared with other state-of-the-art deep learning models where pragmatic inference and flexibility in representing colors from a large continuous space are lacking
Evaluating Subjective Cognitive Appraisals of Emotions from Large Language Models
The emotions we experience involve complex processes; besides physiological
aspects, research in psychology has studied cognitive appraisals where people
assess their situations subjectively, according to their own values (Scherer,
2005). Thus, the same situation can often result in different emotional
experiences. While the detection of emotion is a well-established task, there
is very limited work so far on the automatic prediction of cognitive
appraisals. This work fills the gap by presenting CovidET-Appraisals, the most
comprehensive dataset to-date that assesses 24 appraisal dimensions, each with
a natural language rationale, across 241 Reddit posts. CovidET-Appraisals
presents an ideal testbed to evaluate the ability of large language models --
excelling at a wide range of NLP tasks -- to automatically assess and explain
cognitive appraisals. We found that while the best models are performant,
open-sourced LLMs fall short at this task, presenting a new challenge in the
future development of emotionally intelligent models. We release our dataset at
https://github.com/honglizhan/CovidET-Appraisals-Public.Comment: EMNLP 2023 (Findings) Camera-Ready Versio
Factorized Inference in Deep Markov Models for Incomplete Multimodal Time Series
Integrating deep learning with latent state space models has the potential to
yield temporal models that are powerful, yet tractable and interpretable.
Unfortunately, current models are not designed to handle missing data or
multiple data modalities, which are both prevalent in real-world data. In this
work, we introduce a factorized inference method for Multimodal Deep Markov
Models (MDMMs), allowing us to filter and smooth in the presence of missing
data, while also performing uncertainty-aware multimodal fusion. We derive this
method by factorizing the posterior p(z|x) for non-linear state space models,
and develop a variational backward-forward algorithm for inference. Because our
method handles incompleteness over both time and modalities, it is capable of
interpolation, extrapolation, conditional generation, label prediction, and
weakly supervised learning of multimodal time series. We demonstrate these
capabilities on both synthetic and real-world multimodal data under high levels
of data deletion. Our method performs well even with more than 50% missing
data, and outperforms existing deep approaches to inference in latent time
series.Comment: 8 pages, 4 figures, accepted to AAAI 2020, code available at:
https://github.com/ztangent/multimodal-dm
Improving Multi-Agent Cooperation using Theory of Mind
Recent advances in Artificial Intelligence have produced agents that can beat
human world champions at games like Go, Starcraft, and Dota2. However, most of
these models do not seem to play in a human-like manner: People infer others'
intentions from their behaviour, and use these inferences in scheming and
strategizing. Here, using a Bayesian Theory of Mind (ToM) approach, we
investigated how much an explicit representation of others' intentions improves
performance in a cooperative game. We compared the performance of humans
playing with optimal-planning agents with and without ToM, in a cooperative
game where players have to flexibly cooperate to achieve joint goals. We find
that teams with ToM agents significantly outperform non-ToM agents when
collaborating with all types of partners: non-ToM, ToM, as well as human
players, and that the benefit of ToM increases the more ToM agents there are.
These findings have implications for designing better cooperative agents
Dislocations and Vacancies in Two-Dimensional Mixed Crystals of Spheres and Dimers
In colloidal crystals of spheres, dislocation motion is unrestricted. On the other hand, recent studies of relaxation in crystals of colloidal dimer particles have demonstrated that the dislocation dynamics in such crystals are reminiscent of glassy systems. The observed glassy dynamics arise as a result of dislocation cages formed by certain dimer orientations. In the current study, we use experiments and simulations to investigate the transition that arises when a pure sphere crystal is doped with an increasing concentration of dimers. Specifically, we focus on both dislocation caging and vacancy motion. Interestingly, we find that any nonzero fraction of dimers introduces finite dislocation cages, suggesting that glassy dynamics are present for any mixed crystal. However, we have also identified a vacancy-mediated uncaging mechanism for releasing dislocations from their cages. This mechanism is dependent on vacancy diffusion, which slows by orders of magnitude as the dimer concentration is increased. We propose that in mixed crystals with low dimer concentrations vacancy diffusion is fast enough to uncage dislocations and delay the onset of glassy dislocation dynamics
Large Language Models Produce Responses Perceived to be Empathic
Large Language Models (LLMs) have demonstrated surprising performance on many
tasks, including writing supportive messages that display empathy. Here, we had
these models generate empathic messages in response to posts describing common
life experiences, such as workplace situations, parenting, relationships, and
other anxiety- and anger-eliciting situations. Across two studies (N=192, 202),
we showed human raters a variety of responses written by several models (GPT4
Turbo, Llama2, and Mistral), and had people rate these responses on how
empathic they seemed to be. We found that LLM-generated responses were
consistently rated as more empathic than human-written responses. Linguistic
analyses also show that these models write in distinct, predictable ``styles",
in terms of their use of punctuation, emojis, and certain words. These results
highlight the potential of using LLMs to enhance human peer support in contexts
where empathy is important
- …