Abstract
We consider bounds on the prediction error of classification algorithms based on sample compression. We refine the notion of a compression scheme to distinguish permutation and repetition invariant and non-permutation and repetition invariant compression schemes leading to different prediction error bounds. Also, we extend known results on compression to the case of non-zero empirical risk. We provide bounds on the prediction error of classifiers returned by mistake-driven online learning algorithms by interpreting mistake bounds as bounds on the size of the respective compression scheme of the algorithm. This leads to a bound on the prediction error of perceptron solutions that depends on the margin a support vector machine would achieve on the same training sample. Furthermore, using the property of compression we derive bounds on the average prediction error of kernel classifiers in the PAC-Bayesian framework. These bounds assume a prior measure over the expansion coefficients in the data-dependent kernel expansion and bound the average prediction error uniformly over subsets of the space of expansion coefficients. 2005 Springer Science + Business Media, Inc.
Author supplied keywords
Cite
CITATION STYLE
Graepel, T., Herbrich, R., & Shawe-Taylor, J. (2005). PAC-bayesian compression bounds on the prediction error of learning algorithms for classification. Machine Learning, 59(1–2), 55–76. https://doi.org/10.1007/s10994-005-0462-7
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.