36 research outputs found
Learning the kernel with hyperkernels
This paper addresses the problem of choosing a kernel suitable for estimation with a support
vector machine, hence further automating machine learning. This goal is achieved by defining
a reproducing kernel Hilbert space on the space of kernels itself. Such a formulation leads to a
statistical estimation problem similar to the problem of minimizing a regularized risk functional.
We state the equivalent representer theorem for the choice of kernels and present a semidefinite
programming formulation of the resulting optimization problem. Several recipes for constructing
hyperkernels are provided, as well as the details of common machine learning problems. Experimental
results for classification, regression and novelty detection on UCI data show the feasibility
of our approach
Efficient Output Kernel Learning for Multiple Tasks
The paradigm of multi-task learning is that one can achieve better
generalization by learning tasks jointly and thus exploiting the similarity
between the tasks rather than learning them independently of each other. While
previously the relationship between tasks had to be user-defined in the form of
an output kernel, recent approaches jointly learn the tasks and the output
kernel. As the output kernel is a positive semidefinite matrix, the resulting
optimization problems are not scalable in the number of tasks as an
eigendecomposition is required in each step. \mbox{Using} the theory of
positive semidefinite kernels we show in this paper that for a certain class of
regularizers on the output kernel, the constraint of being positive
semidefinite can be dropped as it is automatically satisfied for the relaxed
problem. This leads to an unconstrained dual problem which can be solved
efficiently. Experiments on several multi-task and multi-class data sets
illustrate the efficacy of our approach in terms of computational efficiency as
well as generalization performance
Regularización Laplaciana en el espacio dual para SVMs
Máster Universitario en en Investigación e Innovación en Inteligencia Computacional y Sistemas InteractivosNowadays, Machine Learning (ML) is a field with a great impact because of its usefulness in solving
many types of problems. However, today large amounts of data are handled and therefore traditional
learning methods can be severely limited in performance. To address this problem, Regularized Learning
(RL) is used, where the objective is to make the model as flexible as possible but preserving the
generalization properties, so that overfitting is avoided.
There are many models that use regularization in their formulations, such as Lasso, or models that
use intrinsic regularization, such as the Support Vector Machine (SVM). In this model, the margin of
a separating hyperplane is maximized, resulting in a solution that depends only on a subset of the
samples called support vectors.
This Master Thesis aims to develop an SVM model with Laplacian regularization in the dual space,
under the intuitive idea that close patterns should have similar coefficients. To construct the Laplacian
term we will use as basis the Fused Lasso model which penalizes the differences of the consecutive
coefficients, but in our case we seek to penalize the differences between every pair of samples, using
the elements of the kernel matrix as weights.
This thesis presents the different phases carried out in the implementation of the new proposal,
starting from the standard SVM, followed by the comparative experiments between the new model and
the original method. As a result, we see that Laplacian regularization is very useful, since the new
proposal outperforms the standard SVM in most of the datasets used, both in classification and regression.
Furthermore, we observe that if we only consider the Laplacian term and we set the parameter
C (upper bound for the coefficients) as if it were infinite, we also obtain better performance than the
standard SVM metho
Regularized System Identification
This open access book provides a comprehensive treatment of recent developments in kernel-based identification that are of interest to anyone engaged in learning dynamic systems from data. The reader is led step by step into understanding of a novel paradigm that leverages the power of machine learning without losing sight of the system-theoretical principles of black-box identification. The authors’ reformulation of the identification problem in the light of regularization theory not only offers new insight on classical questions, but paves the way to new and powerful algorithms for a variety of linear and nonlinear problems. Regression methods such as regularization networks and support vector machines are the basis of techniques that extend the function-estimation problem to the estimation of dynamic models. Many examples, also from real-world applications, illustrate the comparative advantages of the new nonparametric approach with respect to classic parametric prediction error methods. The challenges it addresses lie at the intersection of several disciplines so Regularized System Identification will be of interest to a variety of researchers and practitioners in the areas of control systems, machine learning, statistics, and data science. This is an open access book