4 research outputs found
Harmonizing Global Voices: Culturally-Aware Models for Enhanced Content Moderation
Content moderation at scale faces the challenge of considering local cultural
distinctions when assessing content. While global policies aim to maintain
decision-making consistency and prevent arbitrary rule enforcement, they often
overlook regional variations in interpreting natural language as expressed in
content. In this study, we are looking into how moderation systems can tackle
this issue by adapting to local comprehension nuances. We train large language
models on extensive datasets of media news and articles to create culturally
attuned models. The latter aim to capture the nuances of communication across
geographies with the goal of recognizing cultural and societal variations in
what is considered offensive content. We further explore the capability of
these models to generate explanations for instances of content violation,
aiming to shed light on how policy guidelines are perceived when cultural and
societal contexts change. We find that training on extensive media datasets
successfully induced cultural awareness and resulted in improvements in
handling content violations on a regional basis. Additionally, these
advancements include the ability to provide explanations that align with the
specific local norms and nuances as evidenced by the annotators' preference in
our conducted study. This multifaceted success reinforces the critical role of
an adaptable content moderation approach in keeping pace with the ever-evolving
nature of the content it oversees.Comment: 12 pages, 8 Figures. Supplementary materia
Deception Detection and Rumor Debunking for Social Media
Abstract
The main premise of this chapter is that the time is ripe for more extensive research and development of social media tools that filter out intentionally deceptive information such as deceptive memes, rumors and hoaxes, fake news or other fake posts, tweets and fraudulent profiles. Social media users’ awareness of intentional manipulation of online content appears to be relatively low, while the reliance on unverified information (often obtained from strangers) is at an all-time high. I argue there is need for content verification, systematic fact-checking and filtering of social media streams. This literature survey provides a background for understanding current automated deception detection research, rumor debunking, and broader content verification methodologies, suggests a path towards hybrid technologies, and explains why the development and adoption of such tools might still be a significant challenge