Location of Repository

A low variance error boosting algorithm

By Ching-Wei Wang and Andrew Hunter

Abstract

This paper introduces a robust variant of AdaBoost,\ud cw-AdaBoost, that uses weight perturbation to reduce\ud variance error, and is particularly effective when dealing with data sets, such as microarray data, which have large numbers of features and small number of instances. The algorithm is compared with AdaBoost, Arcing and MultiBoost, using twelve gene expression\ud datasets, using 10-fold cross validation. The new algorithm\ud consistently achieves higher classification accuracy over all these datasets. In contrast to other AdaBoost variants, the algorithm is not susceptible to problems when a zero-error base classifier is encountered

Topics: G700 Artificial Intelligence
Publisher: Springer Netherlands
Year: 2009
OAI identifier: oai:eprints.lincoln.ac.uk:1842

Suggested articles

Preview

Citations

  1. (2000). Additive logistic regression: A statistical view of boosting.
  2. (2001). An Adaptive Version of the Boost by Majority Algorithm.
  3. (1999). An empirical comparison of voting classification algorithms: Bagging, boosting, and variants.
  4. (2004). BagBoosting for tumor classification with gene expression data. Bioinformatics
  5. (1996). Bagging predictors.
  6. (1996). Bagging, boosting and c4.5.
  7. Bias plus variance decomposition for zero-one loss functions.
  8. (2003). Boosting with diverse base classifiers.
  9. (1999). Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays.
  10. (2003). Ensemble machine learning on gene expression data for cancer classification.
  11. (1996). Error Reduction through Learning Multiple Descriptions.
  12. (2002). et al.: Classification, subtype discovery, and prediction of outcome in pediatric acute lymphoblastic leukemia by gene expression profiling.
  13. (2000). et al.: Distinct types of diffuse large B-cell lymphoma identified by gene expression profiling.
  14. (2002). et al.: Gene expression profiling predicts clinical outcome of breast cancer.
  15. (2002). et al.: Genome-wide cDNA microarray screening to correlate gene expression profiles with sensitivity of 85 human cancer xenografts to anticancer drugs.
  16. (2002). et al.: MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia.
  17. (1999). et al.: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring.
  18. (2002). et al.: Prediction of central nervous system embryonal tumour outcome based on gene expression.
  19. (1996). Experiments with a new boosting algorithm.
  20. (2002). Gene expression correlates of clinical prostate cancer behavior.
  21. (2003). Gene expression-based classification of malignant gliomas correlates better with survival than histological classification.
  22. (2005). I.: Diversity in multiple classifier systems.
  23. (2000). MadaBoost: A modification of AdaBoost.
  24. (2000). MultiBoosting: A Technique for Combining Boosting and Wagging.
  25. (2001). Multiple Randomized Classifiers.
  26. (2006). New Ensemble Machine Learning Method for Classification and Prediction on Gene Expression Data.
  27. (2002). Selection bias in gene extraction on the basis of microarray gene-expression data.
  28. (2002). Stochastic Gradient Boosting.
  29. (2006). Totally corrective boosting algorithms that maximize the margin.
  30. (2002). Translation of Microarray Data into Clinically Relevant Cancer Diagnostic Tests Using Gene Expression Ratios in Lung Cancer and Mesothelioma.
  31. (2003). V.B.: Boosting and Microarray Data.
  32. (1996). Variance, and Arcing Classifiers.

To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.