This is an electronic version of the paper presented at the 19th European Symposium on Artificial Neural Networks, held in Bruges on 2011Least-Squares Support Vector Machines (LS-SVMs) have
been successfully applied in many classification and regression tasks. Their
main drawback is the lack of sparseness of the final models. Thus, a
procedure to sparsify LS-SVMs is a frequent desideratum. In this paper,
we adapt to the LS-SVM case a recent work for sparsifying classical SVM
classifiers, which is based on an iterative approximation to the L0-norm.
Experiments on real-world classification and regression datasets illustrate
that this adaptation achieves very sparse models, without significant loss
of accuracy compared to standard LS-SVMs or SVMs