54 research outputs found
Reverse Engineering Socialbot Infiltration Strategies in Twitter
Data extracted from social networks like Twitter are increasingly being used
to build applications and services that mine and summarize public reactions to
events, such as traffic monitoring platforms, identification of epidemic
outbreaks, and public perception about people and brands. However, such
services are vulnerable to attacks from socialbots automated accounts that
mimic real users seeking to tamper statistics by posting messages generated
automatically and interacting with legitimate users. Potentially, if created in
large scale, socialbots could be used to bias or even invalidate many existing
services, by infiltrating the social networks and acquiring trust of other
users with time. This study aims at understanding infiltration strategies of
socialbots in the Twitter microblogging platform. To this end, we create 120
socialbot accounts with different characteristics and strategies (e.g., gender
specified in the profile, how active they are, the method used to generate
their tweets, and the group of users they interact with), and investigate the
extent to which these bots are able to infiltrate the Twitter social network.
Our results show that even socialbots employing simple automated mechanisms are
able to successfully infiltrate the network. Additionally, using a
factorial design, we quantify infiltration effectiveness of different bot
strategies. Our analysis unveils findings that are key for the design of
detection and counter measurements approaches
Do Social Bots Dream of Electric Sheep? A Categorisation of Social Media Bot Accounts
So-called 'social bots' have garnered a lot of attention lately. Previous
research showed that they attempted to influence political events such as the
Brexit referendum and the US presidential elections. It remains, however,
somewhat unclear what exactly can be understood by the term 'social bot'. This
paper addresses the need to better understand the intentions of bots on social
media and to develop a shared understanding of how 'social' bots differ from
other types of bots. We thus describe a systematic review of publications that
researched bot accounts on social media. Based on the results of this
literature review, we propose a scheme for categorising bot accounts on social
media sites. Our scheme groups bot accounts by two dimensions - Imitation of
human behaviour and Intent.Comment: Accepted for publication in the Proceedings of the Australasian
Conference on Information Systems, 201
Online Misinformation: Challenges and Future Directions
Misinformation has become a common part of our digital media environments and it is compromising the ability of our societies to form informed opinions. It generates misperceptions, which have affected the decision making processes in many domains, including economy, health, environment, and elections, among others. Misinformation and its generation, propagation, impact, and management is being studied through a variety of lenses (computer science, social science, journalism, psychology, etc.) since it widely affects multiple aspects of society. In this paper we analyse the phenomenon of misinformation from a technological point of view.We study the current socio-technical advancements towards addressing the problem, identify some of the key limitations of current technologies, and propose some ideas to target such limitations. The goal of this position paper is to reflect on the current state of the art and to stimulate discussions on the future design and development of algorithms, methodologies, and applications
- …