5 research outputs found

    A comparison of pruning criteria for probability trees

    No full text
    Probability trees are decision trees that predict class probabilities rather than the most likely class. The pruning criterion used to learn a probability tree strongly influences the size of the tree and thereby also the quality of its probability estimates. While the effect of pruning criteria on classification accuracy is well-studied, only recently there is more interest in the effect on probability estimates. Hence, it is currently unclear which pruning criteria for probability trees are preferable under which circumstances. In this paper we survey six of the most important pruning criteria for probability trees, and discuss their theoretical advantages and disadvantages. We also perform an extensive experimental study of the relative performance of these pruning criteria. The main conclusion is that overall a pruning criterion based on randomization tests performs best because it is most robust to extreme data characteristics (such as class skew or a high number of classes). We also identify and explain several shortcomings of the other pruning criteria.status: publishe

    A comparison of pruning criteria for probability trees

    No full text
    Probability trees (or Probability Estimation Trees, PET's) are decision trees with probability distributions in the leaves. Usually decision trees are learned in a top-down manner with pre- or postpruning. Which pruning criterion is used strongly influences the size of the resulting tree and the quality of the probability estimates. While the effect of pruning criteria on classification accuracy is well-studied, only recently there is more interest in the effect on probability estimates or probability-based rankings. Hence, there is currently no clear view on the relative performance of all different pruning criteria for probability trees and it is unclear which criteria are preferable under which circumstances. In this paper we survey six of the most important pruning criteria for probability trees. We discuss their theoretical advantages and disadvantages and we perform an extensive experimental study of their relative performance. The main conclusion is that a pruning criterion based on randomization tests usually performs best and learns trees that are relatively small. We identify several scenarios in which other pruning criteria achieve results comparable to those of randomization tests.status: publishe
    corecore