This is an electronic version of the paper presented at the 18th European Symposium on Artificial Neural Networks, held in Bruges on 2010Least Squares Support Vector Machines (LS-SVMs) were
proposed by replacing the inequality constraints inherent to L1-SVMs with
equality constraints. So far this idea has only been suggested for a least
squares (L2) loss. We describe how this can also be done for the sumof-slacks
(L1) loss, yielding a new classifier (Least 1-Norm SVMs) which
gives similar models in terms of complexity and accuracy and that may
also be more robust than LS-SVMs with respect to outliers.With partial support of Spain’s TIN 2007–66862 project and Cátedra IIC en Modelado y
Predicción. The first author is kindly supported by FPU-MICINN grant reference AP2007–
00142