2,097 research outputs found
BRUNO: A Deep Recurrent Model for Exchangeable Data
We present a novel model architecture which leverages deep learning tools to
perform exact Bayesian inference on sets of high dimensional, complex
observations. Our model is provably exchangeable, meaning that the joint
distribution over observations is invariant under permutation: this property
lies at the heart of Bayesian inference. The model does not require variational
approximations to train, and new samples can be generated conditional on
previous samples, with cost linear in the size of the conditioning set. The
advantages of our architecture are demonstrated on learning tasks that require
generalisation from short observed sequences while modelling sequence
variability, such as conditional image generation, few-shot learning, and
anomaly detection.Comment: NIPS 201
Uninformed Students: Student-Teacher Anomaly Detection with Discriminative Latent Embeddings
We introduce a powerful student-teacher framework for the challenging problem
of unsupervised anomaly detection and pixel-precise anomaly segmentation in
high-resolution images. Student networks are trained to regress the output of a
descriptive teacher network that was pretrained on a large dataset of patches
from natural images. This circumvents the need for prior data annotation.
Anomalies are detected when the outputs of the student networks differ from
that of the teacher network. This happens when they fail to generalize outside
the manifold of anomaly-free training data. The intrinsic uncertainty in the
student networks is used as an additional scoring function that indicates
anomalies. We compare our method to a large number of existing deep learning
based methods for unsupervised anomaly detection. Our experiments demonstrate
improvements over state-of-the-art methods on a number of real-world datasets,
including the recently introduced MVTec Anomaly Detection dataset that was
specifically designed to benchmark anomaly segmentation algorithms.Comment: Accepted to CVPR 202
Hybrid Models with Deep and Invertible Features
We propose a neural hybrid model consisting of a linear model defined on a
set of features computed by a deep, invertible transformation (i.e. a
normalizing flow). An attractive property of our model is that both
p(features), the density of the features, and p(targets | features), the
predictive distribution, can be computed exactly in a single feed-forward pass.
We show that our hybrid model, despite the invertibility constraints, achieves
similar accuracy to purely predictive models. Moreover the generative component
remains a good model of the input features despite the hybrid optimization
objective. This offers additional capabilities such as detection of
out-of-distribution inputs and enabling semi-supervised learning. The
availability of the exact joint density p(targets, features) also allows us to
compute many quantities readily, making our hybrid model a useful building
block for downstream applications of probabilistic deep learning.Comment: ICML 201
Sampling - Variational Auto Encoder - Ensemble: In the Quest of Explainable Artificial Intelligence
Explainable Artificial Intelligence (XAI) models have recently attracted a
great deal of interest from a variety of application sectors. Despite
significant developments in this area, there are still no standardized methods
or approaches for understanding AI model outputs. A systematic and cohesive
framework is also increasingly necessary to incorporate new techniques like
discriminative and generative models to close the gap. This paper contributes
to the discourse on XAI by presenting an empirical evaluation based on a novel
framework: Sampling - Variational Auto Encoder (VAE) - Ensemble Anomaly
Detection (SVEAD). It is a hybrid architecture where VAE combined with ensemble
stacking and SHapley Additive exPlanations are used for imbalanced
classification. The finding reveals that combining ensemble stacking, VAE, and
SHAP can. not only lead to better model performance but also provide an easily
explainable framework. This work has used SHAP combined with Permutation
Importance and Individual Conditional Expectations to create a powerful
interpretability of the model. The finding has an important implication in the
real world, where the need for XAI is paramount to boost confidence in AI
applications.Comment: 8 pages, 10 figures, IEEE conference (IEIT 2023
- …