15,607 research outputs found
Sub-Classifier Construction for Error Correcting Output Code Using Minimum Weight Perfect Matching
Multi-class classification is mandatory for real world problems and one of
promising techniques for multi-class classification is Error Correcting Output
Code. We propose a method for constructing the Error Correcting Output Code to
obtain the suitable combination of positive and negative classes encoded to
represent binary classifiers. The minimum weight perfect matching algorithm is
applied to find the optimal pairs of subset of classes by using the
generalization performance as a weighting criterion. Based on our method, each
subset of classes with positive and negative labels is appropriately combined
for learning the binary classifiers. Experimental results show that our
technique gives significantly higher performance compared to traditional
methods including the dense random code and the sparse random code both in
terms of accuracy and classification times. Moreover, our method requires
significantly smaller number of binary classifiers while maintaining accuracy
compared to the One-Versus-One.Comment: 7 pages, 3 figure
Reinforced Decision Trees
In order to speed-up classification models when facing a large number of
categories, one usual approach consists in organizing the categories in a
particular structure, this structure being then used as a way to speed-up the
prediction computation. This is for example the case when using
error-correcting codes or even hierarchies of categories. But in the majority
of approaches, this structure is chosen \textit{by hand}, or during a
preliminary step, and not integrated in the learning process. We propose a new
model called Reinforced Decision Tree which simultaneously learns how to
organize categories in a tree structure and how to classify any input based on
this structure. This approach keeps the advantages of existing techniques (low
inference complexity) but allows one to build efficient classifiers in one
learning step. The learning algorithm is inspired by reinforcement learning and
policy-gradient techniques which allows us to integrate the two steps (building
the tree, and learning the classifier) in one single algorithm
- …