136,357 research outputs found
Tensor-on-tensor regression
We propose a framework for the linear prediction of a multi-way array (i.e.,
a tensor) from another multi-way array of arbitrary dimension, using the
contracted tensor product. This framework generalizes several existing
approaches, including methods to predict a scalar outcome from a tensor, a
matrix from a matrix, or a tensor from a scalar. We describe an approach that
exploits the multiway structure of both the predictors and the outcomes by
restricting the coefficients to have reduced CP-rank. We propose a general and
efficient algorithm for penalized least-squares estimation, which allows for a
ridge (L_2) penalty on the coefficients. The objective is shown to give the
mode of a Bayesian posterior, which motivates a Gibbs sampling algorithm for
inference. We illustrate the approach with an application to facial image data.
An R package is available at https://github.com/lockEF/MultiwayRegression .Comment: 33 pages, 3 figure
Binary Classifier Calibration using an Ensemble of Near Isotonic Regression Models
Learning accurate probabilistic models from data is crucial in many practical
tasks in data mining. In this paper we present a new non-parametric calibration
method called \textit{ensemble of near isotonic regression} (ENIR). The method
can be considered as an extension of BBQ, a recently proposed calibration
method, as well as the commonly used calibration method based on isotonic
regression. ENIR is designed to address the key limitation of isotonic
regression which is the monotonicity assumption of the predictions. Similar to
BBQ, the method post-processes the output of a binary classifier to obtain
calibrated probabilities. Thus it can be combined with many existing
classification models. We demonstrate the performance of ENIR on synthetic and
real datasets for the commonly used binary classification models. Experimental
results show that the method outperforms several common binary classifier
calibration methods. In particular on the real data, ENIR commonly performs
statistically significantly better than the other methods, and never worse. It
is able to improve the calibration power of classifiers, while retaining their
discrimination power. The method is also computationally tractable for large
scale datasets, as it is time, where is the number of
samples
- …