4 research outputs found
W2WNet: A two-module probabilistic Convolutional Neural Network with embedded data cleansing functionality
Ideally, Convolutional Neural Networks (CNNs) should be trained with high quality images with minimum noise and correct ground truth labels. Nonetheless, in many real-world scenarios, such high quality is very hard to obtain, and datasets may be affected by any sort of image degradation and mislabelling issues. This negatively impacts the performance of standard CNNs, both during the training and the inference phase. To
address this issue we propose Wise2WipedNet (W2WNet), a new two-module Convolutional Neural Network, where a Wise module exploits Bayesian inference to identify and discard spurious images during the training and a Wiped module takes care of the final classification, while broadcasting information on the prediction confidence at inference time. The goodness of our solution is demonstrated on a number of public benchmarks addressing different image classification tasks, as well as on a real-world case study on histological image analysis. Overall, our experiments demonstrate that W2WNet is able to identify image degradation and mislabelling issues both at training and at inference time, with positive impact on the final classification accurac
Exploiting generative self-supervised learning for the assessment of biological images with lack of annotations
Computer-aided analysis of biological images typically requires extensive training on large-scale annotated datasets, which is not viable in many situations. In this paper, we present Generative Adversarial Network Discriminator Learner (GAN-DL), a novel self-supervised learning paradigm based on the StyleGAN2 architecture, which we employ for self-supervised image representation learning in the case of fluorescent biological images
Uncertainty-Informed Deep Learning Models Enable High-Confidence Predictions for Digital Histopathology
A model's ability to express its own predictive uncertainty is an essential
attribute for maintaining clinical user confidence as computational biomarkers
are deployed into real-world medical settings. In the domain of cancer digital
histopathology, we describe a novel, clinically-oriented approach to
uncertainty quantification (UQ) for whole-slide images, estimating uncertainty
using dropout and calculating thresholds on training data to establish cutoffs
for low- and high-confidence predictions. We train models to identify lung
adenocarcinoma vs. squamous cell carcinoma and show that high-confidence
predictions outperform predictions without UQ, in both cross-validation and
testing on two large external datasets spanning multiple institutions. Our
testing strategy closely approximates real-world application, with predictions
generated on unsupervised, unannotated slides using predetermined thresholds.
Furthermore, we show that UQ thresholding remains reliable in the setting of
domain shift, with accurate high-confidence predictions of adenocarcinoma vs.
squamous cell carcinoma for out-of-distribution, non-lung cancer cohorts