Abstract
We propose new bounds on the error of learning algorithms in terms of a data-dependent notion of complexity. The estimates we establish give optimal rates and are based on a local and empirical version of Rademacher averages, in the sense that the Rademacher averages are computed from the data, on a subset of functions with small empirical error. We present some applications to classification and prediction with convex function classes, and with kernel classes in particular. © Institute of Mathematical Statistics, 2005.
Author supplied keywords
Cite
CITATION STYLE
Bartlett, P. L., Bousquet, O., & Mendelson, S. (2005). Local rademacher complexities. Annals of Statistics, 33(4), 1497–1537. https://doi.org/10.1214/009053605000000282
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.