10 research outputs found

    Women Are Beautiful, Men Are Leaders: Gender Stereotypes in Machine Translation and Language Modeling

    Full text link
    We present GEST -- a new dataset for measuring gender-stereotypical reasoning in masked LMs and English-to-X machine translation systems. GEST contains samples that are compatible with 9 Slavic languages and English for 16 gender stereotypes about men and women (e.g., Women are beautiful, Men are leaders). The definition of said stereotypes was informed by gender experts. We used GEST to evaluate 11 masked LMs and 4 machine translation systems. We discovered significant and consistent amounts of stereotypical reasoning in almost all the evaluated models and languages

    Disinformation Capabilities of Large Language Models

    Full text link
    Automated disinformation generation is often listed as one of the risks of large language models (LLMs). The theoretical ability to flood the information space with disinformation content might have dramatic consequences for democratic societies around the world. This paper presents a comprehensive study of the disinformation capabilities of the current generation of LLMs to generate false news articles in English language. In our study, we evaluated the capabilities of 10 LLMs using 20 disinformation narratives. We evaluated several aspects of the LLMs: how well they are at generating news articles, how strongly they tend to agree or disagree with the disinformation narratives, how often they generate safety warnings, etc. We also evaluated the abilities of detection models to detect these articles as LLM-generated. We conclude that LLMs are able to generate convincing news articles that agree with dangerous disinformation narratives

    Cross-Lingual Learning With Distributed Representations

    No full text
    Cross-lingual Learning can help to bring state-of-the-art deep learning solutions to smaller languages. These languages in general lack resource for training advanced neural networks. With transfer of knowledge across languages we can improve the results for various NLP tasks

    Učenie s prenosom medzi prirodzenými jazykmi

    Get PDF
    Hlboké učenie sa aktuálne javí ako veľmi perspektívny prístup k mnohým úlohám spracovania prirodzeného jazyka. Tento úspech sa však za-tiaľ prejavuje najmä pri jazykoch, ktoré majú dostatočné množstvo zdrojov na natrénovanie komplexných neurónových sietí. Menšie jazyky s menším množ-stvom zdrojov majú problém tieto nové techniky využiť a priepasť medzi nimi a “bohatými“ jazykmi sa tak prehlbuje. V našej práci sa venujeme tomu, ako túto priepasť zmenšiť pomocou prenosu naučenej informácie z jazyka do jazyka. Hlavnou myšlienkou je trénovanie hlbokých neurónových sietí v multilingvál-nom režime tak, aby sa model naučil využívať znalosti z jedného jazyka aj pre vstupy z iných. Navrhli a vykonali sme experiment s prenosom informácie o sentimente slov pomocou zdieľaného priestoru distribučných vektorov. V ex-perimente sme dosiahli výsledky porovnateľné s nemeckými manuálne vytvo-renými lexikónmi sentimentu, pričom sme však nepoužili žiadne nemecké dáta týkajúce sa sentimentu

    Report on the Current State of Societal Biases in Slovak AI

    No full text
    <p>This report serves as a concluding report of our project <i>Societal Biases in Slovak AI r</i>unning from November 2022 to October 2023. <strong>The goal of our project was to understand how gender biases in particular affect Slovak AI systems</strong>, but also to increase the awareness of this issue in both the expert community and the general public, as this is an issue that affects virtually everyone using modern communication technologies. Our aim was to approach this issue from an interdisciplinary point of view. The team included people with technical expertise in AI and natural language processing; but also AI ethics experts, social scientists, translators, and gender experts.</p><p>The main contribution of our project is the evaluation of multiple types of AI systems, namely English-to-Slovak machine translation systems, Slovak masked language models, and Slovak speech-to-text systems. We have proposed and implemented an evaluation methodology and observed whether these systems exhibit biased behavior with respect to gender. Each experiment was aimed at a specific type of biased behavior, such as male as norm behavior, stereotypical thinking, or non-equitable performance. <strong>Worryingly, we were able to discover some sort of problematic behavior in all the models taken into consideration</strong></p><h3>Contents of the report</h3><ul><li><strong>Chapter 1</strong> introduces the topic, our project and the role of this report.</li><li><strong>Chapter 2 </strong>provides an executive summary of this report and presents the main findings and conclusions of the project. It can be read to ascertain what the main outcomes of our project are at glance.</li><li><strong>Chapter 3</strong> introduces the issue of fairness in AI and provides a brief theoretical background for our research. It introduces and discusses terms such as fairness, bias, equality, algorithmic harms and others.</li><li><strong>Chapter 4 </strong>is an invited chapter written by gender experts Mariana Szapuová and Janka Kottulová. The chapter discusses what gender stereotypes are, how they impact our society and language.</li><li><strong>Chapter 5</strong> describes the experimental research we conducted. It documents how we evaluated the presence of gender biases in various Slovak AI systems, introduces the results, and discusses the implications of what we have found.</li><li><strong>Chapter 6 </strong>documents a data ethics assessment performed during the project to ensure that our research on sensitive issues is as ethical as possible, taking into consideration the needs and points of view of all the stakeholders.</li></ul><p><i>The project is co-funded by the U.S. Embassy Bratislava</i></p><p> </p&gt

    Multilingual Previously Fact-Checked Claim Retrieval

    Full text link
    Fact-checkers are often hampered by the sheer amount of online content that needs to be fact-checked. NLP can help them by retrieving already existing fact-checks relevant to the content being investigated. This paper introduces a new multilingual dataset -- MultiClaim -- for previously fact-checked claim retrieval. We collected 28k posts in 27 languages from social media, 206k fact-checks in 39 languages written by professional fact-checkers, as well as 31k connections between these two groups. This is the most extensive and the most linguistically diverse dataset of this kind to date. We evaluated how different unsupervised methods fare on this dataset and its various dimensions. We show that evaluating such a diverse dataset has its complexities and proper care needs to be taken before interpreting the results. We also evaluated a supervised fine-tuning approach, improving upon the unsupervised method significantly

    Hate Speech Operationalization: A Preliminary Examination of Hate Speech Indicators and Their Structure

    No full text
    Hate speech should be tackled and prosecuted based on how it is operationalized. However, the existing theoretical definitions of hate speech are not sufficiently fleshed out or easily operable. In order to overcome this inadequacy, and with the help of interdisciplinary experts, we propose an empirical definition of hate speech by providing a list of 10 hate speech indicators and the rationale behind them (the indicators refer to specific, observable, and measurable characteristics that offer a practical definition of hate speech). A preliminary exploratory examination of the structure of hate speech, with the focus on comments related to migrants (one of the most reported grounds of hate speech), revealed that two indicators in particular, denial of human rights and promoting violent behavior, occupy a central role in the network of indicators. Furthermore, we discuss the practical implications of the proposed hate speech indicators – especially (semi-)automatic detection using the latest natural language processing (NLP) and machine learning (ML) methods. Having a set of quantifiable indicators could benefit researchers, human right activists, educators, analysts, and regulators by providing them with a pragmatic approach to hate speech assessment and detection

    MultiClaim: Multilingual Previously Fact-Checked Claim Retrieval

    No full text
    MultiClaim: Multilingual Previously Fact-Checked Claim Retrieval - is a dataset that can be used to train a test models used for disinformation combatting. The dataset consists of 206k claims fact-checked by professional fact-checkers and 28k social media posts gathered from the wild. Each social media post has at least on claim assigned. The main idea is to develop information retrieval models that will assign appropriate claims to all the posts. GitHub repository: https://github.com/kinit-sk/multiclaim Contents fact_check_post_mapping.csv - Mapping between fact checks and social media posts: fact_check_id post_id fact_checks.csv - Data about fact-checks: fact_check_id claim - This is the translated text (see below) of the fact-check claim instances - Instances of the fact-check – a list of timestamps and URLs. title - This is the translated text (see below) of the fact-check title posts.csv - Data about social media posts: post_id instances - Instances of the fact-check – a list of timestamps and what were the social media platforms. ocr - This is a list of translated texts (see below) of the OCR transcripts based on the images attached to the post. verdicts - This is a list of verdicts attached by Meta (e.g., False information) text - This is the translated text (see below) of the text written by the user. What is a translated text? A tuple of text, its translation to English and detected languages, e.g., in the sample below we have an original Croatian text, its translation to English and finally the predicted language composition (hbs = Serbo-Croatian): ( '"...bolnice su pune ? ti ina, muk...upravo sada, bolnica Rebro..tragi no sme no', '"...hospitals are full? silence, silence... right now, Rebro hospital... tragically funny', [('hbs', 1.0)] ) More details TB

    MULTITuDE

    No full text
    <p>MULTITuDE is a dataset for multilingual machine-generated text detection benchmark, described in the <a href="https://arxiv.org/abs/2310.13606">EMNLP 2023 conference paper</a>. It consists of 7992 human-written news texts in 11 languages subsampled from <a href="https://github.com/danielvarab/massive-summ">MassiveSumm</a>, accompanied by 66089 texts generated by 8 large language models (by using headlines of news articles). The creation process and scripts for replication/extension are located in a <a href="https://github.com/kinit-sk/mgt-detection-benchmark">GitHub repository</a>.</p><p>If you use this dataset in any publication, project, tool or in any other form, please, cite the <a href="https://arxiv.org/abs/2310.13606">paper</a>.</p><h2>Fields</h2><p>The dataset has the following fields:</p><ul><li>'text' - a text sample,</li><li>'label' - 0 for human-written text, 1 for machine-generated text,</li><li>'multi_label' - a string representing a large language model that generated the text or the string "human" representing a human-written text,</li><li>'split' - a string identifying train or test split of the dataset for the purpose of training and evaluation respectively,</li><li>'language' - the ISO 639-1 language code identifying the language of the given text,</li><li>'length' - word count of the given text,</li><li>'source' - a string identifying the source dataset / news medium of the given text.</li></ul><h2>Statistics (the number of samples)</h2><p><strong>Splits:</strong></p><ul><li>train - 44786</li><li>test - 29295</li></ul><p><strong>Binary labels:</strong></p><ul><li>0 - 7992</li><li>1 - 66089</li></ul><p><strong>Multiclass labels:</strong></p><ul><li>gpt-3.5-turbo -       8300</li><li>gpt-4 -                    8300</li><li>text-davinci-003 -   8297</li><li>alpaca-lora-30b -   8290</li><li>vicuna-13b -          8287</li><li>opt-66b -                8229</li><li>llama-65b -            8229</li><li>opt-iml-max-1.3b - 8157</li><li>human -                 7992</li></ul><p><strong>Languages:</strong></p><ul><li>English (en) - 29460 (train + test)</li><li>Spanish (es) - 11586 (train + test)</li><li>Russian (ru) - 11578 (train + test)</li><li>Dutch (nl) - 2695 (test)</li><li>Catalan (ca) - 2691 (test)</li><li>Czech (cs) - 2689 (test)</li><li>German (de) - 2685 (test)</li><li>Chinese (zh) - 2683 (test)</li><li>Portuguese (pt) - 2673 (test)</li><li>Arabic (ar) - 2673 (test)</li><li>Ukrainian (uk) - 2668 (test)</li></ul&gt
    corecore