5 research outputs found

    Auditing large language models: a three-layered approach

    Full text link
    The emergence of large language models (LLMs) represents a major advance in artificial intelligence (AI) research. However, the widespread use of LLMs is also coupled with significant ethical and social challenges. Previous research has pointed towards auditing as a promising governance mechanism to help ensure that AI systems are designed and deployed in ways that are ethical, legal, and technically robust. However, existing auditing procedures fail to address the governance challenges posed by LLMs, which are adaptable to a wide range of downstream tasks. To help bridge that gap, we offer three contributions in this article. First, we establish the need to develop new auditing procedures that capture the risks posed by LLMs by analysing the affordances and constraints of existing auditing procedures. Second, we outline a blueprint to audit LLMs in feasible and effective ways by drawing on best practices from IT governance and system engineering. Specifically, we propose a three-layered approach, whereby governance audits, model audits, and application audits complement and inform each other. Finally, we discuss the limitations not only of our three-layered approach but also of the prospect of auditing LLMs at all. Ultimately, this article seeks to expand the methodological toolkit available to technology providers and policymakers who wish to analyse and evaluate LLMs from technical, ethical, and legal perspectives.Comment: Preprint, 29 pages, 2 figure

    Ethics-based auditing of automated decision-making systems: considerations, challenges, na

    Get PDF
    Decisions impacting human lives and livelihoods are increasingly being automated. While the use of automated decision-making systems (ADMS) improves efficiency, it is coupled with ethical risks. Previous research has pointed towards ethics-based auditing (EBA) as a promising governance mechanism for managing the ethical risks ADMS pose. However, the affordances and limitations of EBA have yet to be substantiated by empirical research. This thesis seeks to clarify and resolve fundamental questions surrounding EBA. What are the limitations of EBA? How can feasible and effective EBA procedures be designed? These questions are approached on three levels. The conceptual level concerns what EBA is and how it works. The descriptive level focuses on the challenges organisations face when implementing EBA. The applied level concerns how to design EBA procedures that are feasible and effective in practice. This is an integrated thesis, in which the substantive chapters (3–7) are based on published journal articles. Chapter 3 provides a theoretical explanation of how EBA contributes to good governance; Chapter 4 presents new empirical data from a case study of a real-world EBA implementation; Chapter 5 analyses the role of auditing in the proposed EU AI Act; Chapter 6 provides guidance on how to demarcate the material scope of EBA; and Chapter 7 outlines a blueprint for how to audit ADMS with highly general capabilities. My findings suggest that EBA is subject to significant conceptual, technical, and institutional limitations. However, they also indicate that EBA – if properly designed and implemented – helps organisations identify and mitigate some of the ethical risks ADMS pose. I conclude by providing recommendations for how researchers, industry practitioners, auditors, and policymakers can facilitate the emergence of feasible and effective EBA procedures. This thesis thereby serves the purpose of better equipping societies to reap the benefits of ADMS while managing the associated risks
    corecore