2,601 research outputs found

    A crowd-sourcing approach for translations of minority language user-generated content (UGC)

    Get PDF
    Data sparsity is a common problem for machine translation of minority and less-resourced languages. While data collection for standard, grammatical text can be challenging enough, efforts for collection of parallel user-generated content can be even more challenging. In this paper we describe an approach to collecting English↔Irish translations of user-generated content (tweets) that overcomes some of these hurdles. We show how a crowd-sourced data collection campaign, which was tailored to our target audience (the Irish language community), proved successful in gathering data for a niche domain. We also discuss the reliability of crowd-sourcing English↔Irish tweet translations in terms of quality by reporting on a self-rating approach along with qualified reviewer ratings

    Automatically applying a credibility appraisal tool to track vaccination-related communications shared on social media

    Get PDF
    Background: Tools used to appraise the credibility of health information are time-consuming to apply and require context-specific expertise, limiting their use for quickly identifying and mitigating the spread of misinformation as it emerges. Our aim was to estimate the proportion of vaccination-related posts on Twitter are likely to be misinformation, and how unevenly exposure to misinformation was distributed among Twitter users. Methods: Sampling from 144,878 vaccination-related web pages shared on Twitter between January 2017 and March 2018, we used a seven-point checklist adapted from two validated tools to appraise the credibility of a small subset of 474. These were used to train several classifiers (random forest, support vector machines, and a recurrent neural network with transfer learning), using the text from a web page to predict whether the information satisfies each of the seven criteria. Results: Applying the best performing classifier to the 144,878 web pages, we found that 14.4\% of relevant posts to text-based communications were linked to webpages of low credibility and made up 9.2\% of all potential vaccination-related exposures. However, the 100 most popular links to misinformation were potentially seen by between 2 million and 80 million Twitter users, and for a substantial sub-population of Twitter users engaging with vaccination-related information, links to misinformation appear to dominate the vaccination-related information to which they were exposed. Conclusions: We proposed a new method for automatically appraising the credibility of webpages based on a combination of validated checklist tools. The results suggest that an automatic credibility appraisal tool can be used to find populations at higher risk of exposure to misinformation or applied proactively to add friction to the sharing of low credibility vaccination information.Comment: 8 Pages, 5 Figure

    Should I Trust Social Media? How Media Credibility and Language Affect False Memory

    Get PDF
    This study examined the influence of credibility and .language in Internet-based media on false memory. A randomized factorial 2 (media credibility) × 2 (language) experimental design was conducted with 106 college students. The two groups of media credibility consisted of social media (LINE) and non-social media (detik.com), while media language consisted of formal and informal language. A confidence test was used to measure false memory. A two-factor ANOVA showed that media credibility significantly affects false memory. Participants in the detik.com group were more confident in the information received and had greater false memory than the LINE group. However, no significant effect of language was found, and no significant interaction effect between media credibility and language on false memory was found. This study suggests that individuals should be cautious when reading information on non-social media platforms, as individuals tend to place more confidence on the source, leading to greater false memory

    How many others have shared this? Experimentally investigating the effects of social cues on engagement, misinformation, and unpredictability on social media

    Full text link
    Unlike traditional media, social media typically provides quantified metrics of how many users have engaged with each piece of content. Some have argued that the presence of these cues promotes the spread of misinformation. Here we investigate the causal effect of social cues on users' engagement with social media posts. We conducted an experiment with N=628 Americans on a custom-built newsfeed interface where we systematically varied the presence and strength of social cues. We find that when cues are shown, indicating that a larger number of others have engaged with a post, users were more likely to share and like that post. Furthermore, relative to a control without social cues, the presence of social cues increased the sharing of true relative to false news. The presence of social cues also makes it more difficult to precisely predict how popular any given post would be. Together, our results suggest that -- instead of distracting users or causing them to share low-quality news -- social cues may, in certain circumstances, actually boost truth discernment and reduce the sharing of misinformation. Our work suggests that social cues play important roles in shaping users' attention and engagement on social media, and platforms should understand the effects of different cues before making changes to what cues are displayed and how
    corecore