Bounding the generalization error of convex combinations of classifiers: Balancing the dimensionality and the margins

12Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

A problem of bounding the generalization error of a classifier f ∈ conv(H), where H is a "base" class of functions (classifiers), is considered. This problem frequently occurs in computer learning, where efficient algorithms that combine simple classifiers into a complex one (such as boosting and bagging) have attracted a lot of attention. Using Talagrand's concentration inequalities for empirical processes, we obtain new sharper bounds on the generalization error of combined classifiers that take into account both the empirical distribution of "classification margins" and an "approximate dimension" of the classifiers, and study the performance of these bounds in several experiments with learning algorithms.

Cite

CITATION STYLE

APA

Koltchinskii, V., Panchenko, D., & Lozano, F. (2003). Bounding the generalization error of convex combinations of classifiers: Balancing the dimensionality and the margins. Annals of Applied Probability, 13(1), 213–252. https://doi.org/10.1214/aoap/1042765667

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free