9,128 research outputs found

    People Talking and AI Listening: How Stigmatizing Language in EHR Notes Affect AI Performance

    Full text link
    Electronic health records (EHRs) serve as an essential data source for the envisioned artificial intelligence (AI)-driven transformation in healthcare. However, clinician biases reflected in EHR notes can lead to AI models inheriting and amplifying these biases, perpetuating health disparities. This study investigates the impact of stigmatizing language (SL) in EHR notes on mortality prediction using a Transformer-based deep learning model and explainable AI (XAI) techniques. Our findings demonstrate that SL written by clinicians adversely affects AI performance, particularly so for black patients, highlighting SL as a source of racial disparity in AI model development. To explore an operationally efficient way to mitigate SL's impact, we investigate patterns in the generation of SL through a clinicians' collaborative network, identifying central clinicians as having a stronger impact on racial disparity in the AI model. We find that removing SL written by central clinicians is a more efficient bias reduction strategy than eliminating all SL in the entire corpus of data. This study provides actionable insights for responsible AI development and contributes to understanding clinician behavior and EHR note writing in healthcare.Comment: 54 pages, 9 figure

    The Relations Between Pedagogical and Scientific Explanations of Algorithms: Case Studies from the French Administration

    Get PDF
    The opacity of some recent Machine Learning (ML) techniques have raised fundamental questions on their explainability, and created a whole domain dedicated to Explainable Artificial Intelligence (XAI). However, most of the literature has been dedicated to explainability as a scientific problem dealt with typical methods of computer science, from statistics to UX. In this paper, we focus on explainability as a pedagogical problem emerging from the interaction between lay users and complex technological systems. We defend an empirical methodology based on field work, which should go beyond the in-vitro analysis of UX to examine in-vivo problems emerging in the field. Our methodology is also comparative, as it chooses to steer away from the almost exclusive focus on ML to compare its challenges with those faced by more vintage algorithms. Finally, it is also philosophical, as we defend the relevance of the philosophical literature to define the epistemic desiderata of a good explanation. This study was conducted in collaboration with Etalab, a Task Force of the French Prime Minister in charge of Open Data & Open Government Policies, dealing in particular with the enforcement of the right to an explanation. In order to illustrate and refine our methodology before going up to scale, we conduct a preliminary work of case studies on the main different types of algorithms used by the French administration: computation, matching algorithms and ML. We study the merits and drawbacks of a recent approach to explanation, which we baptize input-output black box reasoning or BBR for short. We begin by presenting a conceptual framework including the distinctions necessary to a study of pedagogical explainability. We proceed to algorithmic case studies, and draw model-specific and model-agnostic lessons and conjectures

    Possibilities and implications of using the ICF and other vocabulary standards in electronic health records

    Get PDF
    There is now widespread recognition of the powerful potential of electronic health record (EHR) systems to improve the health-care delivery system. The benefits of EHRs grow even larger when the health data within their purview are seamlessly shared, aggregated and processed across different providers, settings and institutions. Yet, the plethora of idiosyncratic conventions for identifying the same clinical content in different information systems is a fundamental barrier to fully leveraging the potential of EHRs. Only by adopting vocabulary standards that provide the lingua franca across these local dialects can computers efficiently move, aggregate and use health data for decision support, outcomes management, quality reporting, research and many other purposes. In this regard, the International Classification of Functioning, Disability, and Health (ICF) is an important standard for physiotherapists because it provides a framework and standard language for describing health and health-related states. However, physiotherapists and other health-care professionals capture a wide range of data such as patient histories, clinical findings, tests and measurements, procedures, and so on, for which other vocabulary standards such as Logical Observation Identifiers Names and Codes and Systematized Nomenclature Of Medicine Clinical Terms are crucial for interoperable communication between different electronic systems. In this paper, we describe how the ICF and other internationally accepted vocabulary standards could advance physiotherapy practise and research by enabling data sharing and reuse by EHRs. We highlight how these different vocabulary standards fit together within a comprehensive record system, and how EHRs can make use of them, with a particular focus on enhancing decision-making. By incorporating the ICF and other internationally accepted vocabulary standards into our clinical information systems, physiotherapists will be able to leverage the potent capabilities of EHRs and contribute our unique clinical perspective to other health-care providers within the emerging electronic health information infrastructure

    The Care2Report System: Automated Medical Reporting as an Integrated Solution to Reduce Administrative Burden in Healthcare

    Get PDF
    Documenting patient medical information in the electronic medical record is a time-consuming task at the expense of direct patient care. We propose an integrated solution to automate the process of medical reporting. This vision is enabled through the integration of speech and action recognition technology with semantic interpretation based on knowledge graphs. This paper presents our dialogue summarization pipeline that transforms speech into a medical report via transcription and formal representation. We discuss the functional and technical architecture of our Care2Report system along with an initial system evaluation with data of real consultation sessions

    Clinical information extraction for preterm birth risk prediction

    Get PDF
    This paper contributes to the pursuit of leveraging unstructured medical notes to structured clinical decision making. In particular, we present a pipeline for clinical information extraction from medical notes related to preterm birth, and discuss the main challenges as well as its potential for clinical practice. A large collection of medical notes, created by staff during hospitalizations of patients who were at risk of delivering preterm, was gathered and analyzed. Based on an annotated collection of notes, we trained and evaluated information extraction components to discover clinical entities such as symptoms, events, anatomical sites and procedures, as well as attributes linked to these clinical entities. In a retrospective study, we show that these are highly informative for clinical decision support models that are trained to predict whether delivery is likely to occur within specific time windows, in combination with structured information from electronic health records
    corecore