14 research outputs found

    Popular Ensemble Methods: An Empirical Study

    Full text link
    An ensemble consists of a set of individually trained classifiers (such as neural networks or decision trees) whose predictions are combined when classifying novel instances. Previous research has shown that an ensemble is often more accurate than any of the single classifiers in the ensemble. Bagging (Breiman, 1996c) and Boosting (Freund and Shapire, 1996; Shapire, 1990) are two relatively new but popular methods for producing ensembles. In this paper we evaluate these methods on 23 data sets using both neural networks and decision trees as our classification algorithm. Our results clearly indicate a number of conclusions. First, while Bagging is almost always more accurate than a single classifier, it is sometimes much less accurate than Boosting. On the other hand, Boosting can create ensembles that are less accurate than a single classifier -- especially when using neural networks. Analysis indicates that the performance of the Boosting methods is dependent on the characteristics of the data set being examined. In fact, further results show that Boosting ensembles may overfit noisy data sets, thus decreasing its performance. Finally, consistent with previous studies, our work suggests that most of the gain in an ensemble's performance comes in the first few classifiers combined; however, relatively large gains can be seen up to 25 classifiers when Boosting decision trees

    Forgetting Exceptions is Harmful in Language Learning

    Get PDF
    We show that in language learning, contrary to received wisdom, keeping exceptional training instances in memory can be beneficial for generalization accuracy. We investigate this phenomenon empirically on a selection of benchmark natural language processing tasks: grapheme-to-phoneme conversion, part-of-speech tagging, prepositional-phrase attachment, and base noun phrase chunking. In a first series of experiments we combine memory-based learning with training set editing techniques, in which instances are edited based on their typicality and class prediction strength. Results show that editing exceptional instances (with low typicality or low class prediction strength) tends to harm generalization accuracy. In a second series of experiments we compare memory-based learning and decision-tree learning methods on the same selection of tasks, and find that decision-tree learning often performs worse than memory-based learning. Moreover, the decrease in performance can be linked to the degree of abstraction from exceptions (i.e., pruning or eagerness). We provide explanations for both results in terms of the properties of the natural language processing tasks and the learning algorithms.Comment: 31 pages, 7 figures, 10 tables. uses 11pt, fullname, a4wide tex styles. Pre-print version of article to appear in Machine Learning 11:1-3, Special Issue on Natural Language Learning. Figures on page 22 slightly compressed to avoid page overloa

    Noise-tolerant inverse analysis models for nondestructive evaluation of transportation infrastructure systems using neural networks

    Get PDF
    The need to rapidly and cost-effectively evaluate the present condition of pavement infrastructure is a critical issue concerning the deterioration of ageing transportation infrastructure all around the world. Nondestructive testing (NDT) and evaluation methods are well-suited for characterising materials and determining structural integrity of pavement systems. The falling weight deflectometer (FWD) is a NDT equipment used to assess the structural condition of highway and airfield pavement systems and to determine the moduli of pavement layers. This involves static or dynamic inverse analysis (referred to as backcalculation) of FWD deflection profiles in the pavement surface under a simulated truck load. The main objective of this study was to employ biologically inspired computational systems to develop robust pavement layer moduli backcalculation algorithms that can tolerate noise or inaccuracies in the FWD deflection data collected in the field. Artificial neural systems, also known as artificial neural networks (ANNs), are valuable computational intelligence tools that are increasingly being used to solve resource-intensive complex engineering problems. Unlike the linear elastic layered theory commonly used in pavement layer backcalculation, non-linear unbound aggregate base and subgrade soil response models were used in an axisymmetric finite element structural analysis programme to generate synthetic database for training and testing the ANN models. In order to develop more robust networks that can tolerate the noisy or inaccurate pavement deflection patterns in the NDT data, several network architectures were trained with varying levels of noise in them. The trained ANN models were capable of rapidly predicting the pavement layer moduli and critical pavement responses (tensile strains at the bottom of the asphalt concrete layer, compressive strains on top of the subgrade layer and the deviator stresses on top of the subgrade layer), and also pavement surface deflections with very low average errors comparable with those obtained directly from the finite element analyses
    corecore