1 research outputs found
Learning a hyperplane classifier by minimizing an exact bound on the VC dimension
The VC dimension measures the capacity of a learning machine, and a low VC
dimension leads to good generalization. While SVMs produce state-of-the-art
learning performance, it is well known that the VC dimension of a SVM can be
unbounded; despite good results in practice, there is no guarantee of good
generalization. In this paper, we show how to learn a hyperplane classifier by
minimizing an exact, or \boldmath{} bound on its VC dimension. The
proposed approach, termed as the Minimal Complexity Machine (MCM), involves
solving a simple linear programming problem. Experimental results show, that on
a number of benchmark datasets, the proposed approach learns classifiers with
error rates much less than conventional SVMs, while often using fewer support
vectors. On many benchmark datasets, the number of support vectors is less than
one-tenth the number used by SVMs, indicating that the MCM does indeed learn
simpler representations.Comment: Accepted Author Manuscript (Neurocomputing, Elsevier); 10 page