61 research outputs found
LIMEcraft: Handcrafted superpixel selection and inspection for Visual eXplanations
The increased interest in deep learning applications, and their
hard-to-detect biases result in the need to validate and explain complex
models. However, current explanation methods are limited as far as both the
explanation of the reasoning process and prediction results are concerned. They
usually only show the location in the image that was important for model
prediction. The lack of possibility to interact with explanations makes it
difficult to verify and understand exactly how the model works. This creates a
significant risk when using the model. The risk is compounded by the fact that
explanations do not take into account the semantic meaning of the explained
objects. To escape from the trap of static and meaningless explanations, we
propose a tool and a process called LIMEcraft. LIMEcraft enhances the process
of explanation by allowing a user to interactively select semantically
consistent areas and thoroughly examine the prediction for the image instance
in case of many image features. Experiments on several models show that our
tool improves model safety by inspecting model fairness for image pieces that
may indicate model bias. The code is available at:
http://github.com/MI2DataLab/LIMEcraf
Hospital Length of Stay Prediction Based on Multi-modal Data towards Trustworthy Human-AI Collaboration in Radiomics
To what extent can the patient's length of stay in a hospital be predicted
using only an X-ray image? We answer this question by comparing the performance
of machine learning survival models on a novel multi-modal dataset created from
1235 images with textual radiology reports annotated by humans. Although
black-box models predict better on average than interpretable ones, like Cox
proportional hazards, they are not inherently understandable. To overcome this
trust issue, we introduce time-dependent model explanations into the human-AI
decision making process. Explaining models built on both: human-annotated and
algorithm-extracted radiomics features provides valuable insights for
physicians working in a hospital. We believe the presented approach to be
general and widely applicable to other time-to-event medical use cases. For
reproducibility, we open-source code and the TLOS dataset at
https://github.com/mi2datalab/xlungs-trustworthy-los-prediction.Comment: Accepted at International Conference on Artificial Intelligence in
Medicine (AIME 2023
Towards Evaluating Explanations of Vision Transformers for Medical Imaging
As deep learning models increasingly find applications in critical domains
such as medical imaging, the need for transparent and trustworthy
decision-making becomes paramount. Many explainability methods provide insights
into how these models make predictions by attributing importance to input
features. As Vision Transformer (ViT) becomes a promising alternative to
convolutional neural networks for image classification, its interpretability
remains an open research question. This paper investigates the performance of
various interpretation methods on a ViT applied to classify chest X-ray images.
We introduce the notion of evaluating faithfulness, sensitivity, and complexity
of ViT explanations. The obtained results indicate that Layerwise relevance
propagation for transformers outperforms Local interpretable model-agnostic
explanations and Attention visualization, providing a more accurate and
reliable representation of what a ViT has actually learned. Our findings
provide insights into the applicability of ViT explanations in medical imaging
and highlight the importance of using appropriate evaluation criteria for
comparing them.Comment: Accepted by XAI4CV Workshop at CVPR 202
Multi-task learning for classification, segmentation, reconstruction, and detection on chest CT scans
Lung cancer and covid-19 have one of the highest morbidity and mortality
rates in the world. For physicians, the identification of lesions is difficult
in the early stages of the disease and time-consuming. Therefore, multi-task
learning is an approach to extracting important features, such as lesions, from
small amounts of medical data because it learns to generalize better. We
propose a novel multi-task framework for classification, segmentation,
reconstruction, and detection. To the best of our knowledge, we are the first
ones who added detection to the multi-task solution. Additionally, we checked
the possibility of using two different backbones and different loss functions
in the segmentation task.Comment: presented at the Polish Conference on Artificial Intelligence
(PP-RAI), 202
SurvSHAP(t): Time-dependent explanations of machine learning survival models
Machine and deep learning survival models demonstrate similar or even
improved time-to-event prediction capabilities compared to classical
statistical learning methods yet are too complex to be interpreted by humans.
Several model-agnostic explanations are available to overcome this issue;
however, none directly explain the survival function prediction. In this paper,
we introduce SurvSHAP(t), the first time-dependent explanation that allows for
interpreting survival black-box models. It is based on SHapley Additive
exPlanations with solid theoretical foundations and a broad adoption among
machine learning practitioners. The proposed methods aim to enhance precision
diagnostics and support domain experts in making decisions. Experiments on
synthetic and medical data confirm that SurvSHAP(t) can detect variables with a
time-dependent effect, and its aggregation is a better determinant of the
importance of variables for a prediction than SurvLIME. SurvSHAP(t) is
model-agnostic and can be applied to all models with functional output. We
provide an accessible implementation of time-dependent explanations in Python
at http://github.com/MI2DataLab/survshap
intsvy: An R Package for Analyzing International Large-Scale Assessment Data
This paper introduces intsvy, an R package for working with international assessment data (e.g., PISA, TIMSS, PIRLS). The package includes functions for importing data, performing data analysis, and visualizing results. The paper describes the underlying methodology and provides real data examples. Tools for importing data allow useRs to select variables from student, home, school, and teacher survey instruments as well as for specific countries. Data analysis functions take into account the complex sample design (with replicate weights) and rotated test forms (with plausible values of achievement scores) in the calculation of point estimates and standard errors of means, standard deviations, regression coefficients, correlation coefficients, and frequency tables. Visualization tools present data aggregates in standardized graphical form
- …