556 research outputs found

    Interpretability and Explainability: A Machine Learning Zoo Mini-tour

    Full text link
    In this review, we examine the problem of designing interpretable and explainable machine learning models. Interpretability and explainability lie at the core of many machine learning and statistical applications in medicine, economics, law, and natural sciences. Although interpretability and explainability have escaped a clear universal definition, many techniques motivated by these properties have been developed over the recent 30 years with the focus currently shifting towards deep learning methods. In this review, we emphasise the divide between interpretability and explainability and illustrate these two different research directions with concrete examples of the state-of-the-art. The review is intended for a general machine learning audience with interest in exploring the problems of interpretation and explanation beyond logistic regression or random forest variable importance. This work is not an exhaustive literature survey, but rather a primer focusing selectively on certain lines of research which the authors found interesting or informative

    (Un)reasonable Allure of Ante-hoc Interpretability for High-stakes Domains: Transparency Is Necessary but Insufficient for Explainability

    Full text link
    Ante-hoc interpretability has become the holy grail of explainable machine learning for high-stakes domains such as healthcare; however, this notion is elusive, lacks a widely-accepted definition and depends on the deployment context. It can refer to predictive models whose structure adheres to domain-specific constraints, or ones that are inherently transparent. The latter notion assumes observers who judge this quality, whereas the former presupposes them to have technical and domain expertise, in certain cases rendering such models unintelligible. Additionally, its distinction from the less desirable post-hoc explainability, which refers to methods that construct a separate explanatory model, is vague given that transparent predictors may still require (post-)processing to yield satisfactory explanatory insights. Ante-hoc interpretability is thus an overloaded concept that comprises a range of implicit properties, which we unpack in this paper to better understand what is needed for its safe deployment across high-stakes domains. To this end, we outline model- and explainer-specific desiderata that allow us to navigate its distinct realisations in view of the envisaged application and audience

    Doing et undoing gender dans les crĂšches: une analyse desinteractions des Ă©ducateurs/Ă©ducatrices avec les enfants

    Full text link
    Kinderkrippen sind fĂŒr viele Kinder die erste Bildungsinstitution und fĂŒr die Gleichstellung der Geschlechter bedeutsam. Im Beitrag wird auf der Basis einer ethnographischen Videostudie in vier Deutschschweizer Kinderkrippen untersucht, wie Gender in der pĂ€dagogischen Alltagspraxis der Kinderbetreuenden relevant wird. FĂŒr die Kodierung der Videodaten werden InteraktionsverlĂ€ufe in Bezug auf doing und undoing gender, Dramatisierung und Dethematisierung analysiert. Die Ergebnisse zeigen, dass die Kinderbetreuenden das von Kindern gezeigte Verhalten, sei es doing oder undoing gender, verstĂ€rken, jedoch selten intervenieren um Gleichstellung herzustellen. Zur Förderung der Gleichstellung in der Kita sind die Organisationskultur und die pĂ€dagogische QualitĂ€t entscheidend. (DIPF/Orig.)Pour nombreux enfants, la crĂšche est la premiĂšre institution Ă©ducative qu’ils frĂ©quentent. La crĂšche donc est importante par rapport Ă  la question de l’éducation Ă  l’égalitĂ© des sexes. Dans cet article, sur la base d’une Ă©tude ethnographique et de donnĂ©es vidĂ©o effectuĂ©es dans quatre crĂšches, nous examinons comment les questions de genre s’actualisent dans les pratiques pĂ©dagogiques quotidiennes. Le codage des interactions a Ă©tĂ© rĂ©alisĂ© Ă  partir des concepts de doing gender, undoing gender, dramatisation et dĂ©-thĂ©matisation. Les rĂ©sultats montrent que les Ă©ducateurs et Ă©ducatrices renforcent les comportements des enfants et exigent rarement l’égalitĂ©. Pour soutenir la question de l’éducation Ă  l’égalitĂ© dans les crĂšches, la culture d’organisation et la qualitĂ© de l’éducation s’avĂšrent ĂȘtre primordiales. (DIPF/Orig.

    Generation of Differentially Private Heterogeneous Electronic Health Records

    Full text link
    Electronic Health Records (EHRs) are commonly used by the machine learning community for research on problems specifically related to health care and medicine. EHRs have the advantages that they can be easily distributed and contain many features useful for e.g. classification problems. What makes EHR data sets different from typical machine learning data sets is that they are often very sparse, due to their high dimensionality, and often contain heterogeneous (mixed) data types. Furthermore, the data sets deal with sensitive information, which limits the distribution of any models learned using them, due to privacy concerns. For these reasons, using EHR data in practice presents a real challenge. In this work, we explore using Generative Adversarial Networks to generate synthetic, heterogeneous EHRs with the goal of using these synthetic records in place of existing data sets for downstream classification tasks. We will further explore applying differential privacy (DP) preserving optimization in order to produce DP synthetic EHR data sets, which provide rigorous privacy guarantees, and are therefore shareable and usable in the real world. The performance (measured by AUROC, AUPRC and accuracy) of our model's synthetic, heterogeneous data is very close to the original data set (within 3 - 5% of the baseline) for the non-DP model when tested in a binary classification task. Using strong (1,10−5)(1, 10^{-5}) DP, our model still produces data useful for machine learning tasks, albeit incurring a roughly 17% performance penalty in our tested classification task. We additionally perform a sub-population analysis and find that our model does not introduce any bias into the synthetic EHR data compared to the baseline in either male/female populations, or the 0-18, 19-50 and 51+ age groups in terms of classification performance for either the non-DP or DP variant
    • 

    corecore