One of the surprising recurring phenomena observed in experiments
with boosting is that the test error of the generated classifier usually does
not increase as its size becomes very large, and often is observed to decrease
even after the training error reaches zero. In this paper, we show that this
phenomenon is related to the distribution of margins of the training
examples with respect to the generated voting classification rule, where the
margin of an example is simply the difference between the number of correct
votes and the maximum number of votes received by any incorrect label. We show
that techniques used in the analysis of Vapnik’s support vector
classifiers and of neural networks with small weights can be applied to voting
methods to relate the margin distribution to the test error. We also show
theoretically and experimentally that boosting is especially effective at
increasing the margins of the training examples. Finally, we compare our
explanation to those based on the bias-variance decomposition.
@article{1024691352,
author = {Schapire, Robert E. and Freund, Yoav and Bartlett, Peter and Lee, Wee Sun},
title = {Boosting the margin: a new explanation for the effectiveness of
voting methods},
journal = {Ann. Statist.},
volume = {26},
number = {3},
year = {1998},
pages = { 1651-1686},
language = {en},
url = {http://dml.mathdoc.fr/item/1024691352}
}
Schapire, Robert E.; Freund, Yoav; Bartlett, Peter; Lee, Wee Sun. Boosting the margin: a new explanation for the effectiveness of
voting methods. Ann. Statist., Tome 26 (1998) no. 3, pp. 1651-1686. http://gdmltest.u-ga.fr/item/1024691352/