Abstract
This paper describes a new technique for solving multiclass learning problems by combining Freund and Schapire's boosting algorithm with the main ideas of Dietterich and Bakiri's method of error-correcting output codes (ECOC). Boosting is a general method of improving the accuracy of a given base or "weak" learning algorithm. ECOC is a robust method of solving multiclass learning problems by reducing to a sequence of two-class problems. We show that our new hybrid method has advantages of both: Like ECOC, our method only requires that the base learning algorithm work on binary-labeled data. Like boosting, we prove that the method comes with strong theoretical guarantees on the training and generalization error of the final combined hypothesis assuming only that the base learning algorithm perform slightly better than random guessing. Although previous methods were known for boosting multiclass problems, the new method may be significantly faster and require less programming effort in creating the base \nlearning algorithm. We also compare the new algorithm \nexperimentally to other voting methods.
Keywords
Affiliated Institutions
Related Publications
Experiments with a new boosting algorithm
In an earlier paper, we introduced a new "boosting" algorithm called AdaBoost which, theoretically, can be used to significantly reduce the error of any learni...
Solving Multiclass Learning Problems via Error-Correcting Output Codes
Multiclass learning problems involve finding a definitionfor an unknown function f(x) whose range is a discrete setcontaining k > 2 values (i.e., k ``classes''). Thedefinitio...
Machine Learning Bias, Statistical Bias, and Statistical Variance of Decision Tree Algorithms
The term "bias" is widely used---and with different meanings---in the fields of machine learning and statistics. This paper clarifies the uses of this term and...
Improved boosting algorithms using confidence-rated predictions
We describe several improvements to Freund and Schapire‘s AdaBoost boosting algorithm, particularly in a setting in which hypotheses may assign confidences to each of their pred...
Generalized Bradley-Terry Models and Multi-Class Probability Estimates
The Bradley-Terry model for obtaining individual skill from paired comparisons has been popular in many areas. In machine learning, this model is related to multi-class probabil...
Publication Info
- Year
- 1997
- Type
- article
- Pages
- 313-321
- Citations
- 263
- Access
- Closed