2,058 research outputs found
The (de)biasing effect of GAN-based augmentation methods on skin lesion images
New medical datasets are now more open to the public, allowing for better and
more extensive research. Although prepared with the utmost care, new datasets
might still be a source of spurious correlations that affect the learning
process. Moreover, data collections are usually not large enough and are often
unbalanced. One approach to alleviate the data imbalance is using data
augmentation with Generative Adversarial Networks (GANs) to extend the dataset
with high-quality images. GANs are usually trained on the same biased datasets
as the target data, resulting in more biased instances. This work explored
unconditional and conditional GANs to compare their bias inheritance and how
the synthetic data influenced the models. We provided extensive manual data
annotation of possibly biasing artifacts on the well-known ISIC dataset with
skin lesions. In addition, we examined classification models trained on both
real and synthetic data with counterfactual bias explanations. Our experiments
showed that GANs inherited biases and sometimes even amplified them, leading to
even stronger spurious correlations. Manual data annotation and synthetic
images are publicly available for reproducible scientific research.Comment: Accepted to MICCAI202
Targeted Data Augmentation for bias mitigation
The development of fair and ethical AI systems requires careful consideration
of bias mitigation, an area often overlooked or ignored. In this study, we
introduce a novel and efficient approach for addressing biases called Targeted
Data Augmentation (TDA), which leverages classical data augmentation techniques
to tackle the pressing issue of bias in data and models. Unlike the laborious
task of removing biases, our method proposes to insert biases instead,
resulting in improved performance. To identify biases, we annotated two diverse
datasets: a dataset of clinical skin lesions and a dataset of male and female
faces. These bias annotations are published for the first time in this study,
providing a valuable resource for future research. Through Counterfactual Bias
Insertion, we discovered that biases associated with the frame, ruler, and
glasses had a significant impact on models. By randomly introducing biases
during training, we mitigated these biases and achieved a substantial decrease
in bias measures, ranging from two-fold to more than 50-fold, while maintaining
a negligible increase in the error rate
EPVT: Environment-aware Prompt Vision Transformer for Domain Generalization in Skin Lesion Recognition
Skin lesion recognition using deep learning has made remarkable progress, and
there is an increasing need for deploying these systems in real-world
scenarios. However, recent research has revealed that deep neural networks for
skin lesion recognition may overly depend on disease-irrelevant image artifacts
(i.e. dark corners, dense hairs), leading to poor generalization in unseen
environments. To address this issue, we propose a novel domain generalization
method called EPVT, which involves embedding prompts into the vision
transformer to collaboratively learn knowledge from diverse domains.
Concretely, EPVT leverages a set of domain prompts, each of which plays as a
domain expert, to capture domain-specific knowledge; and a shared prompt for
general knowledge over the entire dataset. To facilitate knowledge sharing and
the interaction of different prompts, we introduce a domain prompt generator
that enables low-rank multiplicative updates between domain prompts and the
shared prompt. A domain mixup strategy is additionally devised to reduce the
co-occurring artifacts in each domain, which allows for more flexible decision
margins and mitigates the issue of incorrectly assigned domain labels.
Experiments on four out-of-distribution datasets and six different biased ISIC
datasets demonstrate the superior generalization ability of EPVT in skin lesion
recognition across various environments. Our code and dataset will be released
at https://github.com/SiyuanYan1/EPVT.Comment: 12 pages, 5 figure
- …