377 research outputs found
Towards Measuring Adversarial Twitter Interactions against Candidates in the US Midterm Elections
Adversarial interactions against politicians on social media such as Twitter
have significant impact on society. In particular they disrupt substantive
political discussions online, and may discourage people from seeking public
office. In this study, we measure the adversarial interactions against
candidates for the US House of Representatives during the run-up to the 2018 US
general election. We gather a new dataset consisting of 1.7 million tweets
involving candidates, one of the largest corpora focusing on political
discourse. We then develop a new technique for detecting tweets with toxic
content that are directed at any specific candidate.Such technique allows us to
more accurately quantify adversarial interactions towards political candidates.
Further, we introduce an algorithm to induce candidate-specific adversarial
terms to capture more nuanced adversarial interactions that previous techniques
may not consider toxic. Finally, we use these techniques to outline the breadth
of adversarial interactions seen in the election, including offensive
name-calling, threats of violence, posting discrediting information, attacks on
identity, and adversarial message repetition
Detecting Harmful Agendas in News Articles
Manipulated news online is a growing problem which necessitates the use of
automated systems to curtail its spread. We argue that while misinformation and
disinformation detection have been studied, there has been a lack of investment
in the important open challenge of detecting harmful agendas in news articles;
identifying harmful agendas is critical to flag news campaigns with the
greatest potential for real world harm. Moreover, due to real concerns around
censorship, harmful agenda detectors must be interpretable to be effective. In
this work, we propose this new task and release a dataset, NewsAgendas, of
annotated news articles for agenda identification. We show how interpretable
systems can be effective on this task and demonstrate that they can perform
comparably to black-box models.Comment: Camera-ready for ACL-WASSA 202
Detecting and Characterizing Political Incivility on Social Media
Researchers of political communication study the impact and perceptions of
political incivility on social media. Yet, so far, relatively few works
attempted to automatically detect and characterize political incivility. In our
work, we study political incivility in Twitter, presenting several research
contributions. First, we present state-of-the-art incivility detection results
using a large dataset, which we collected and labeled via crowd sourcing.
Importantly, we distinguish between uncivil political speech that is impolite
and intolerant anti-democratic discourse. Applying political incivility
detection at large-scale, we derive insights regarding the prevalence of this
phenomenon across users, and explore the network characteristics of users who
are susceptible to disseminating uncivil political content online. Finally, we
propose an approach for modeling social context information about the tweet
author alongside the tweet content, showing that this leads to significantly
improved performance on the task of political incivility detection. This result
holds promise for related tasks, such as hate speech and stance detection
Characterizing and Detecting Hateful Users on Twitter
Most current approaches to characterize and detect hate speech focus on
\textit{content} posted in Online Social Networks. They face shortcomings to
collect and annotate hateful speech due to the incompleteness and noisiness of
OSN text and the subjectivity of hate speech. These limitations are often aided
with constraints that oversimplify the problem, such as considering only tweets
containing hate-related words. In this work we partially address these issues
by shifting the focus towards \textit{users}. We develop and employ a robust
methodology to collect and annotate hateful users which does not depend
directly on lexicon and where the users are annotated given their entire
profile. This results in a sample of Twitter's retweet graph containing
users, out of which were annotated. We also collect the users
who were banned in the three months that followed the data collection. We show
that hateful users differ from normal ones in terms of their activity patterns,
word usage and as well as network structure. We obtain similar results
comparing the neighbors of hateful vs. neighbors of normal users and also
suspended users vs. active users, increasing the robustness of our analysis. We
observe that hateful users are densely connected, and thus formulate the hate
speech detection problem as a task of semi-supervised learning over a graph,
exploiting the network of connections on Twitter. We find that a node embedding
algorithm, which exploits the graph structure, outperforms content-based
approaches for the detection of both hateful ( AUC vs AUC) and
suspended users ( AUC vs AUC). Altogether, we present a
user-centric view of hate speech, paving the way for better detection and
understanding of this relevant and challenging issue.Comment: This is an extended version of the homonymous short paper to be
presented at ICWSM-18. arXiv admin note: text overlap with arXiv:1801.0031
Toxicity in the Decentralized Web and the Potential for Model Sharing
The "Decentralised Web" (DW) is an evolving concept, which encompasses technologies aimed at providing greater transparency and openness on the web. The DW relies on independent servers (aka instances) that mesh together in a peer-to-peer fashion to deliver a range of services (e.g. micro-blogs, image sharing, video streaming). However, toxic content moderation in this decentralised context is challenging. This is because there is no central entity that can define toxicity, nor a large central pool of data that can be used to build universal classifiers. It is therefore unsurprising that there have been several high-profile cases of the DW being misused to coordinate and disseminate harmful material. Using a dataset of 9.9M posts from 117K users on Pleroma (a popular DW microblogging service), we quantify the presence of toxic content. We find that toxic content is prevalent and spreads rapidly between instances. We show that automating per-instance content moderation is challenging due to the lack of sufficient training data available and the effort required in labelling. We therefore propose and evaluate ModPair, a model sharing system that effectively detects toxic content, gaining an average per-instance macro-F1 score 0.89
- …