694 research outputs found
Tracking Fluctuations in Psychological States Using Social Media Language: A Case Study of Weekly Emotion
Personality psychologists are increasingly documenting dynamic, withinâperson processes. Big data methodologies can augment this endeavour by allowing for the collection of naturalistic and personalityârelevant digital traces from online environments. Whereas big data methods have primarily been used to catalogue static personality dimensions, here we present a case study in how they can be used to track dynamic fluctuations in psychological states. We apply a textâbased, machine learning prediction model to Facebook status updates to compute weekly trajectories of emotional valence and arousal. We train this model on 2895 humanâannotated Facebook statuses and apply the resulting model to 303 575 Facebook statuses posted by 640 US Facebook users who had previously selfâreported their Big Five traits, yielding an average of 28 weekly estimates per user. We examine the correlations between modelâpredicted emotion and selfâreported personality, providing a test of the robustness of these links when using weekly aggregated data, rather than momentary data as in prior work. We further present dynamic visualizations of weekly valence and arousal for every user, while making the final data set of 17 937 weeks openly available. We discuss the strengths and drawbacks of this method in the context of personality psychologyâs evolution into a dynamic science. © 2020 European Association of Personality PsychologyPeer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/163564/3/per2261-sup-0001-Open_Practices_Disclosure_Form.pdfhttp://deepblue.lib.umich.edu/bitstream/2027.42/163564/2/per2261.pdfhttp://deepblue.lib.umich.edu/bitstream/2027.42/163564/1/per2261_am.pd
An emotional mess! Deciding on a framework for building a Dutch emotion-annotated corpus
Seeing the myriad of existing emotion models, with the categorical versus dimensional opposition the most important dividing line, building an emotion-annotated corpus requires some well thought-out strategies concerning framework choice. In our work on automatic emotion detection in Dutch texts, we investigate this problem by means of two case studies. We find that the labels joy, love, anger, sadness and fear are well-suited to annotate texts coming from various domains and topics, but that the connotation of the labels strongly depends on the origin of the texts. Moreover, it seems that information is lost when an emotional state is forcedly classified in a limited set of categories, indicating that a bi-representational format is desirable when creating an emotion corpus.Seeing the myriad of existing emotion models, with the categorical versus dimensional opposition the most important dividing line, building an emotion-annotated corpus requires some well thought-out strategies concerning framework choice. In our work on automatic emotion detection in Dutch texts, we investigate this problem by means of two case studies. We find that the labels joy, love, anger, sadness and fear are well-suited to annotate texts coming from various domains and topics, but that the connotation of the labels strongly depends on the origin of the texts. Moreover, it seems that information is lost when an emotional state is forcedly classified in a limited set of categories, indicating that a bi-representational format is desirable when creating an emotion corpus.P
Ruddit: Norms of Offensiveness for English Reddit Comments
On social media platforms, hateful and offensive language negatively impact
the mental well-being of users and the participation of people from diverse
backgrounds. Automatic methods to detect offensive language have largely relied
on datasets with categorical labels. However, comments can vary in their degree
of offensiveness. We create the first dataset of English language Reddit
comments that has fine-grained, real-valued scores between -1 (maximally
supportive) and 1 (maximally offensive). The dataset was annotated using
Best--Worst Scaling, a form of comparative annotation that has been shown to
alleviate known biases of using rating scales. We show that the method produces
highly reliable offensiveness scores. Finally, we evaluate the ability of
widely-used neural models to predict offensiveness scores on this new dataset.Comment: Camera-ready version in ACL 202
Introduction to the second international symposium of platial information science
People âliveâ and constitute places every day through recurrent practices and experience. Our everyday lives, however, are complex, and so are places. In contrast to abstract space, the way people experience places includes a range of aspects like physical setting, meaning, and emotional attachment. This inherent complexity requires researchers to investigate the concept of place from a variety of viewpoints. The formal representation of place â a major goal in GIScience related to place â is no exception and can only be successfully addressed if we consider geographical, psychological, anthropological, sociological, cognitive, and other perspectives. This yearâs symposium brings together place-based researchers from different disciplines to discuss the current state of platial research. Therefore, this volume contains contributions from a range of fields including geography, psychology, cognitive science, linguistics, and cartography
- âŠ