4,749 research outputs found
Efficient Multi-Template Learning for Structured Prediction
Conditional random field (CRF) and Structural Support Vector Machine
(Structural SVM) are two state-of-the-art methods for structured prediction
which captures the interdependencies among output variables. The success of
these methods is attributed to the fact that their discriminative models are
able to account for overlapping features on the whole input observations. These
features are usually generated by applying a given set of templates on labeled
data, but improper templates may lead to degraded performance. To alleviate
this issue, in this paper, we propose a novel multiple template learning
paradigm to learn structured prediction and the importance of each template
simultaneously, so that hundreds of arbitrary templates could be added into the
learning model without caution. This paradigm can be formulated as a special
multiple kernel learning problem with exponential number of constraints. Then
we introduce an efficient cutting plane algorithm to solve this problem in the
primal, and its convergence is presented. We also evaluate the proposed
learning paradigm on two widely-studied structured prediction tasks,
\emph{i.e.} sequence labeling and dependency parsing. Extensive experimental
results show that the proposed method outperforms CRFs and Structural SVMs due
to exploiting the importance of each template. Our complexity analysis and
empirical results also show that our proposed method is more efficient than
OnlineMKL on very sparse and high-dimensional data. We further extend this
paradigm for structured prediction using generalized -block norm
regularization with , and experiments show competitive performances when
A Feature Selection Method for Multivariate Performance Measures
Feature selection with specific multivariate performance measures is the key
to the success of many applications, such as image retrieval and text
classification. The existing feature selection methods are usually designed for
classification error. In this paper, we propose a generalized sparse
regularizer. Based on the proposed regularizer, we present a unified feature
selection framework for general loss functions. In particular, we study the
novel feature selection paradigm by optimizing multivariate performance
measures. The resultant formulation is a challenging problem for
high-dimensional data. Hence, a two-layer cutting plane algorithm is proposed
to solve this problem, and the convergence is presented. In addition, we adapt
the proposed method to optimize multivariate measures for multiple instance
learning problems. The analyses by comparing with the state-of-the-art feature
selection methods show that the proposed method is superior to others.
Extensive experiments on large-scale and high-dimensional real world datasets
show that the proposed method outperforms -SVM and SVM-RFE when choosing a
small subset of features, and achieves significantly improved performances over
SVM in terms of -score
Learning with Augmented Features for Heterogeneous Domain Adaptation
We propose a new learning method for heterogeneous domain adaptation (HDA),
in which the data from the source domain and the target domain are represented
by heterogeneous features with different dimensions. Using two different
projection matrices, we first transform the data from two domains into a common
subspace in order to measure the similarity between the data from two domains.
We then propose two new feature mapping functions to augment the transformed
data with their original features and zeros. The existing learning methods
(e.g., SVM and SVR) can be readily incorporated with our newly proposed
augmented feature representations to effectively utilize the data from both
domains for HDA. Using the hinge loss function in SVM as an example, we
introduce the detailed objective function in our method called Heterogeneous
Feature Augmentation (HFA) for a linear case and also describe its
kernelization in order to efficiently cope with the data with very high
dimensions. Moreover, we also develop an alternating optimization algorithm to
effectively solve the nontrivial optimization problem in our HFA method.
Comprehensive experiments on two benchmark datasets clearly demonstrate that
HFA outperforms the existing HDA methods.Comment: ICML201
The Radiance In Your Eyes
https://digitalcommons.library.umaine.edu/mmb-vp/6698/thumbnail.jp
- …