34,950 research outputs found
Analyzing the Targets of Hate in Online Social Media
Social media systems allow Internet users a congenial platform to freely
express their thoughts and opinions. Although this property represents
incredible and unique communication opportunities, it also brings along
important challenges. Online hate speech is an archetypal example of such
challenges. Despite its magnitude and scale, there is a significant gap in
understanding the nature of hate speech on social media. In this paper, we
provide the first of a kind systematic large scale measurement study of the
main targets of hate speech in online social media. To do that, we gather
traces from two social media systems: Whisper and Twitter. We then develop and
validate a methodology to identify hate speech on both these systems. Our
results identify online hate speech forms and offer a broader understanding of
the phenomenon, providing directions for prevention and detection approaches.Comment: Short paper, 4 pages, 4 table
Kek, Cucks, and God Emperor Trump: A Measurement Study of 4chan's Politically Incorrect Forum and its Effects on the Web
The discussion-board site 4chan has been part of the Internet's dark underbelly since its inception, and recent political events have put it increasingly in the spotlight. In particular, /pol/, the “Politically Incorrect'” board, has been a central figure in the outlandish 2016 US election season, as it has often been linked to the alt-right movement and its rhetoric of hate and racism. However, 4chan remains relatively unstudied by the scientific community: little is known about its user base, the content it generates, and how it affects other parts of the Web. In this paper, we start addressing this gap by analyzing /pol/ along several axes, using a dataset of over 8M posts we collected over two and a half months. First, we perform a general characterization, showing that /pol/ users are well distributed around the world and that 4chan's unique features encourage fresh discussions. We also analyze content, finding, for instance, that YouTube links and hate speech are predominant on /pol/. Overall, our analysis not only provides the first measurement study of /pol/, but also insight into online harassment and hate speech trends in social media
Deep-Cov19-Hate: A Textual-Based Novel Approach for Automatic Detection of Hate Speech in Online Social Networks throughout COVID-19 with Shallow and Deep Learning Models
The use of various online social media platforms rising day by day caused an increase in the correct or incorrect information shared by users, especially during COVID-19. The introduction of COVID-19 on the world agenda gave rise to an overall bad reaction against East Asia (esp. China) in online social media platforms. The social media users who spread degrading, racist, disrespectful, abusive, discriminatory, critical, abuse, harsh, offensive, etc. posts accused the Asian people of being responsible for the outbreak of COVID-19. For this reason, the development of the Hate Speech Detection (HSD) system was necessary in order to prevent the spread of these posts about COVID-19. In this article, a textual-based study on COVID-19-related hate speech (HS) sharing in online social networks was carried out with Shallow Learning (SL) and Deep Learning (DL) methods. In the first step of this study, typical Natural Language Processing (NLP) pipeline was applied for gathered two different datasets. This NLP pipeline was performed using bag of words, term frequency, document matrix, etc. techniques for features extraction representing datasets. Then, ten different SL and DL models were fine-tuned for HS datasets related to COVID-19. Accuracy, precision, sensitivity, and F-score performance measurement criteria were calculated to compare the performance of the SL and DL algorithms for the problem of HSD. The RNN, one of the models proposed for the first and second dataset in HSD, prevailed with the highest accuracy values of 78.7% and 90.3%, respectively. Due to the promising results of all approaches operated in the HSD, they are forecasted to be chosen in the solution of many other social media and network problems related to COVID-19
How do you feel? Measuring User-Perceived Value for Rejecting Machine Decisions in Hate Speech Detection
Hate speech moderation remains a challenging task for social media platforms.
Human-AI collaborative systems offer the potential to combine the strengths of
humans' reliability and the scalability of machine learning to tackle this
issue effectively. While methods for task handover in human-AI collaboration
exist that consider the costs of incorrect predictions, insufficient attention
has been paid to accurately estimating these costs. In this work, we propose a
value-sensitive rejection mechanism that automatically rejects machine
decisions for human moderation based on users' value perceptions regarding
machine decisions. We conduct a crowdsourced survey study with 160 participants
to evaluate their perception of correct and incorrect machine decisions in the
domain of hate speech detection, as well as occurrences where the system
rejects making a prediction. Here, we introduce Magnitude Estimation, an
unbounded scale, as the preferred method for measuring user (dis)agreement with
machine decisions. Our results show that Magnitude Estimation can provide a
reliable measurement of participants' perception of machine decisions. By
integrating user-perceived value into human-AI collaboration, we further show
that it can guide us in 1) determining when to accept or reject machine
decisions to obtain the optimal total value a model can deliver and 2)
selecting better classification models as compared to the more widely used
target of model accuracy.Comment: To appear at AIES '23. Philippe Lammerts, Philip Lippmann, Yen-Chia
Hsu, Fabio Casati, and Jie Yang. 2023. How do you feel? Measuring
User-Perceived Value for Rejecting Machine Decisions in Hate Speech
Detection. In AAAI/ACM Conference on AI, Ethics, and Society (AIES '23),
August 8.10, 2023, Montreal, QC, Canada. ACM, New York, NY, USA. 11 page
Thou shalt not hate: Countering Online Hate Speech
Hate content in social media is ever-increasing. While Facebook, Twitter,
Google have attempted to take several steps to tackle the hateful content, they
have mostly been unsuccessful. Counterspeech is seen as an effective way of
tackling the online hate without any harm to the freedom of speech. Thus, an
alternative strategy for these platforms could be to promote counterspeech as a
defense against hate content. However, in order to have a successful promotion
of such counterspeech, one has to have a deep understanding of its dynamics in
the online world. Lack of carefully curated data largely inhibits such
understanding. In this paper, we create and release the first ever dataset for
counterspeech using comments from YouTube. The data contains 13,924 manually
annotated comments where the labels indicate whether a comment is a
counterspeech or not. This data allows us to perform a rigorous measurement
study characterizing the linguistic structure of counterspeech for the first
time. This analysis results in various interesting insights such as: the
counterspeech comments receive much more likes as compared to the
non-counterspeech comments, for certain communities majority of the
non-counterspeech comments tend to be hate speech, the different types of
counterspeech are not all equally effective and the language choice of users
posting counterspeech is largely different from those posting non-counterspeech
as revealed by a detailed psycholinguistic analysis. Finally, we build a set of
machine learning models that are able to automatically detect counterspeech in
YouTube videos with an F1-score of 0.71. We also build multilabel models that
can detect different types of counterspeech in a comment with an F1-score of
0.60.Comment: Accepted at ICWSM 2019. 12 Pages, 5 Figures, and 7 Tables. The
dataset and models are available here:
https://github.com/binny-mathew/Countering_Hate_Speech_ICWSM201
A Quantitative Approach to Understanding Online Antisemitism
A new wave of growing antisemitism, driven by fringe Web communities, is an
increasingly worrying presence in the socio-political realm. The ubiquitous and
global nature of the Web has provided tools used by these groups to spread
their ideology to the rest of the Internet. Although the study of antisemitism
and hate is not new, the scale and rate of change of online data has impacted
the efficacy of traditional approaches to measure and understand these
troubling trends. In this paper, we present a large-scale, quantitative study
of online antisemitism. We collect hundreds of million posts and images from
alt-right Web communities like 4chan's Politically Incorrect board (/pol/) and
Gab. Using scientifically grounded methods, we quantify the escalation and
spread of antisemitic memes and rhetoric across the Web. We find the frequency
of antisemitic content greatly increases (in some cases more than doubling)
after major political events such as the 2016 US Presidential Election and the
"Unite the Right" rally in Charlottesville. We extract semantic embeddings from
our corpus of posts and demonstrate how automated techniques can discover and
categorize the use of antisemitic terminology. We additionally examine the
prevalence and spread of the antisemitic "Happy Merchant" meme, and in
particular how these fringe communities influence its propagation to more
mainstream communities like Twitter and Reddit. Taken together, our results
provide a data-driven, quantitative framework for understanding online
antisemitism. Our methods serve as a framework to augment current qualitative
efforts by anti-hate groups, providing new insights into the growth and spread
of hate online.Comment: To appear at the 14th International AAAI Conference on Web and Social
Media (ICWSM 2020). Please cite accordingl
- …