8 research outputs found

    Hierarchical linear support vector machine

    Full text link
    This is the author’s version of a work that was accepted for publication in Pattern Recognition. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Pattern Recognition, Vol. 45, Iss. 12, (2012) DOI: 10.1016/j.patcog.2012.06.002The increasing size and dimensionality of real-world datasets make it necessary to design efficient algorithms not only in the training process but also in the prediction phase. In applications such as credit card fraud detection, the classifier needs to predict an event in 10 ms at most. In these environments the speed of the prediction constraints heavily outweighs the training costs. We propose a new classification method, called a Hierarchical Linear Support Vector Machine (H-LSVM), based on the construction of an oblique decision tree in which the node split is obtained as a Linear Support Vector Machine. Although other methods have been proposed to break the data space down in subregions to speed up Support Vector Machines, the H-LSVM algorithm represents a very simple and efficient model in training but mainly in prediction for large-scale datasets. Only a few hyperplanes need to be evaluated in the prediction step, no kernel computation is required and the tree structure makes parallelization possible. In experiments with medium and large datasets, the H-LSVM reduces the prediction cost considerably while achieving classification results closer to the non-linear SVM than that of the linear case.The authors would like to thank the anonymous reviewers for their comments that help improve the manuscript. I.R.-L. is supported by an FPU Grant from Universidad Autónoma de Madrid, and partially supported by the Universidad Autónoma de Madrid-IIC Chair and TIN2010-21575-C02-01. R.H. acknowledges partial support by ONRN00014-07-1-0741, USARIEM-W81XWH-10-C-0040 (ELINTRIX) and JPL-2012-1455933

    Cost-Sensitive Boosting

    Full text link

    Cost-sensitive learning and decision making for Massachusetts PIP claim fraud data

    No full text
    For the first time in NFL history, the winning coach of the Super Bowl will be African-American. While the history of African-American coaches in the NFL dates back to Frederick Douglass Fritz Pollard in 1921, the next African-American head coach of an NFL team was not hired until 1989, when Art Shell was hired by the Oakland Raiders. In 2002, the National Football League adopted the “Rooney Rule,” which requires any team engaged in hiring a head coach to assemble a diverse candidate slate. At least in part as a result, there were a record seven black head coaches in the 2006 season, two of whom--Tony Dungy of the Indianapolis Colts and Lovie Smith of the Chicago Bears--are now the first African-American head coaches to lead their teams to the Super Bowl. Additionally, with the appointment of Jerry Reese as General Manager of the New York Giants, there are now a record six African Americans in General Manager or General Manager equivalent positions. ACS\u27s panel, which took place in the official Super Bowl XLI media center in Miami, Florida, addressed how those policies were adopted and implemented, the impact they have had on the league and the challenges that remain

    Cost-sensitive learning and decision making for Massachusetts PIP claim fraud data

    No full text
    In this article, we investigate the issue of cost-sensitive classification for a data set of Massachusetts closed personal injury protection (PIP) automobile insurance claims that were previously investigated for suspicion of fraud by domain experts and for which we obtained cost information. After a theoretical exposition on cost-sensitive learning and decision-making methods, we then apply these methods to the claims data at hand to contrast the predictive performance of the documented methods for a selection of decision tree and rule learners. We use standard logistic regression and (smoothed) naive Bayes as benchmarks. (C) 2004 Wiley Periodicals, Inc.status: publishe

    Minimizing Recommended Error Costs Under Noisy Inputs in Rule-Based Expert Systems

    Get PDF
    This dissertation develops methods to minimize recommendation error costs when inputs to a rule-based expert system are prone to errors. The problem often arises in web-based applications where data are inherently noisy or provided by users who perceive some benefit from falsifying inputs. Prior studies proposed methods that attempted to minimize the probability of recommendation error, but did not take into account the relative costs of different types of errors. In situations where these differences are significant, an approach that minimizes the expected misclassification error costs has advantages over extant methods that ignore these costs. Building on the existing literature, two new techniques - Cost-Based Input Modification (CBIM) and Cost-Based Knowledge-Base Modification (CBKM) were developed and evaluated. Each method takes as inputs (1) the joint probability distribution of a set of rules, (2) the distortion matrix for input noise as characterized by the probability distribution of the observed input vectors conditioned on their true values, and (3) the misclassification cost for each type of recommendation error. Under CBIM, for any observed input vector v, the recommendation is based on a modified input vector v\u27 such that the expected error costs are minimized. Under CBKM the rule base itself is modified to minimize the expected cost of error. The proposed methods were investigated as follows: as a control, in the special case where the costs associated with different types of errors are identical, the recommendations under these methods were compared for consistency with those obtained under extant methods. Next, the relative advantages of CBIM and CBKM were compared as (1) the noise level changed, and (2) the structure of the cost matrix varied. As expected, CBKM and CBIM outperformed the extant Knowledge Base Modification (KM) and Input Modification (IM) methods over a wide range of input distortion and cost matrices, with some restrictions. Under the control, with constant misclassification costs, the new methods performed equally with the extant methods. As misclassification costs increased, CBKM outperformed KM and CBIM outperformed IM. Using different cost matrices to increase misclassification cost asymmetry and order, CBKM and CBIM performance increased. At very low distortion levels, CBKM and CBIM underperformed as error probability became more significant in each method\u27s estimation. Additionally, CBKM outperformed CBIM over a wide range of input distortion as its technique of modifying an original knowledge base outperformed the technique of modifying inputs to an unmodified decision tree
    corecore