24,448 research outputs found
Dropout Inference in Bayesian Neural Networks with Alpha-divergences
To obtain uncertainty estimates with real-world Bayesian deep learning
models, practical inference approximations are needed. Dropout variational
inference (VI) for example has been used for machine vision and medical
applications, but VI can severely underestimates model uncertainty.
Alpha-divergences are alternative divergences to VI's KL objective, which are
able to avoid VI's uncertainty underestimation. But these are hard to use in
practice: existing techniques can only use Gaussian approximating
distributions, and require existing models to be changed radically, thus are of
limited use for practitioners. We propose a re-parametrisation of the
alpha-divergence objectives, deriving a simple inference technique which,
together with dropout, can be easily implemented with existing models by simply
changing the loss of the model. We demonstrate improved uncertainty estimates
and accuracy compared to VI in dropout networks. We study our model's epistemic
uncertainty far away from the data using adversarial images, showing that these
can be distinguished from non-adversarial images by examining our model's
uncertainty
Efficient Active Learning for Image Classification and Segmentation using a Sample Selection and Conditional Generative Adversarial Network
Training robust deep learning (DL) systems for medical image classification
or segmentation is challenging due to limited images covering different disease
types and severity. We propose an active learning (AL) framework to select most
informative samples and add to the training data. We use conditional generative
adversarial networks (cGANs) to generate realistic chest xray images with
different disease characteristics by conditioning its generation on a real
image sample. Informative samples to add to the training set are identified
using a Bayesian neural network. Experiments show our proposed AL framework is
able to achieve state of the art performance by using about 35% of the full
dataset, thus saving significant time and effort over conventional methods
- …