34,376 research outputs found
Deep Dictionary Learning: A PARametric NETwork Approach
Deep dictionary learning seeks multiple dictionaries at different image
scales to capture complementary coherent characteristics. We propose a method
for learning a hierarchy of synthesis dictionaries with an image classification
goal. The dictionaries and classification parameters are trained by a
classification objective, and the sparse features are extracted by reducing a
reconstruction loss in each layer. The reconstruction objectives in some sense
regularize the classification problem and inject source signal information in
the extracted features. The performance of the proposed hierarchical method
increases by adding more layers, which consequently makes this model easier to
tune and adapt. The proposed algorithm furthermore, shows remarkably lower
fooling rate in presence of adversarial perturbation. The validation of the
proposed approach is based on its classification performance using four
benchmark datasets and is compared to a CNN of similar size
High-Dimensional Feature Selection by Feature-Wise Kernelized Lasso
The goal of supervised feature selection is to find a subset of input
features that are responsible for predicting output values. The least absolute
shrinkage and selection operator (Lasso) allows computationally efficient
feature selection based on linear dependency between input features and output
values. In this paper, we consider a feature-wise kernelized Lasso for
capturing non-linear input-output dependency. We first show that, with
particular choices of kernel functions, non-redundant features with strong
statistical dependence on output values can be found in terms of kernel-based
independence measures. We then show that the globally optimal solution can be
efficiently computed; this makes the approach scalable to high-dimensional
problems. The effectiveness of the proposed method is demonstrated through
feature selection experiments with thousands of features.Comment: 18 page
Large-scale Nonlinear Variable Selection via Kernel Random Features
We propose a new method for input variable selection in nonlinear regression.
The method is embedded into a kernel regression machine that can model general
nonlinear functions, not being a priori limited to additive models. This is the
first kernel-based variable selection method applicable to large datasets. It
sidesteps the typical poor scaling properties of kernel methods by mapping the
inputs into a relatively low-dimensional space of random features. The
algorithm discovers the variables relevant for the regression task together
with learning the prediction model through learning the appropriate nonlinear
random feature maps. We demonstrate the outstanding performance of our method
on a set of large-scale synthetic and real datasets.Comment: Final version for proceedings of ECML/PKDD 201
- …