4 research outputs found
On-Off Pattern Encoding and Path-Count Encoding as Deep Neural Network Representations
Understanding the encoded representation of Deep Neural Networks (DNNs) has
been a fundamental yet challenging objective. In this work, we focus on two
possible directions for analyzing representations of DNNs by studying simple
image classification tasks. Specifically, we consider \textit{On-Off pattern}
and \textit{PathCount} for investigating how information is stored in deep
representations. On-off pattern of a neuron is decided as `on' or `off'
depending on whether the neuron's activation after ReLU is non-zero or zero.
PathCount is the number of paths that transmit non-zero energy from the input
to a neuron. We investigate how neurons in the network encodes information by
replacing each layer's activation with On-Off pattern or PathCount and
evaluating its effect on classification performance. We also examine
correlation between representation and PathCount. Finally, we show a possible
way to improve an existing DNN interpretation method, Class Activation Map
(CAM), by directly utilizing On-Off or PathCount.Comment: 8 pages, 4 figure
Martingale Posterior Neural Processes
A Neural Process (NP) estimates a stochastic process implicitly defined with
neural networks given a stream of data, rather than pre-specifying priors
already known, such as Gaussian processes. An ideal NP would learn everything
from data without any inductive biases, but in practice, we often restrict the
class of stochastic processes for the ease of estimation. One such restriction
is the use of a finite-dimensional latent variable accounting for the
uncertainty in the functions drawn from NPs. Some recent works show that this
can be improved with more "data-driven" source of uncertainty such as
bootstrapping. In this work, we take a different approach based on the
martingale posterior, a recently developed alternative to Bayesian inference.
For the martingale posterior, instead of specifying prior-likelihood pairs, a
predictive distribution for future data is specified. Under specific conditions
on the predictive distribution, it can be shown that the uncertainty in the
generated future data actually corresponds to the uncertainty of the implicitly
defined Bayesian posteriors. Based on this result, instead of assuming any form
of the latent variables, we equip a NP with a predictive distribution
implicitly defined with neural networks and use the corresponding martingale
posteriors as the source of uncertainty. The resulting model, which we name as
Martingale Posterior Neural Process (MPNP), is demonstrated to outperform
baselines on various tasks.Comment: ICLR 202
Probabilistic Imputation for Time-series Classification with Missing Data
Multivariate time series data for real-world applications typically contain a
significant amount of missing values. The dominant approach for classification
with such missing values is to impute them heuristically with specific values
(zero, mean, values of adjacent time-steps) or learnable parameters. However,
these simple strategies do not take the data generative process into account,
and more importantly, do not effectively capture the uncertainty in prediction
due to the multiple possibilities for the missing values. In this paper, we
propose a novel probabilistic framework for classification with multivariate
time series data with missing values. Our model consists of two parts; a deep
generative model for missing value imputation and a classifier. Extending the
existing deep generative models to better capture structures of time-series
data, our deep generative model part is trained to impute the missing values in
multiple plausible ways, effectively modeling the uncertainty of the
imputation. The classifier part takes the time series data along with the
imputed missing values and classifies signals, and is trained to capture the
predictive uncertainty due to the multiple possibilities of imputations.
Importantly, we show that na\"ively combining the generative model and the
classifier could result in trivial solutions where the generative model does
not produce meaningful imputations. To resolve this, we present a novel
regularization technique that can promote the model to produce useful
imputation values that help classification. Through extensive experiments on
real-world time series data with missing values, we demonstrate the
effectiveness of our method