Data dependent priors in PAC-Bayes bounds

1Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

One of the central aims of Statistical Learning Theory is the bounding of the test set performance of classifiers trained with i.i.d. data. For Support Vector Machines the tightest technique for assessing this so-called generalisation error is known as the PAC-Bayes theorem. The bound holds independently of the choice of prior, but better priors lead to sharper bounds. The priors leading to the tightest bounds to date are spherical Gaussian distributions whose means are determined from a separate subset of data. This paper gives another turn of the screw by introducing a further data dependence on the shape of the prior: The separate data set determines a direction along which the covariance matrix of the prior is stretched in order to sharpen the bound. In addition, we present a classification algorithm that aims at minimizing the bound as a design criterion and whose generalisation can be easily analysed in terms of the new bound. The experimental work includes a set of classification tasks preceded by a bound-driven model selection. These experiments illustrate how the new bound acting on the new classifier can be much tighter than the original PAC-Bayes Bound applied to an SVM, and lead to more accurate classifiers. © Springer-Verlag Berlin Heidelberg 2010.

Cite

CITATION STYLE

APA

Shawe- Taylor, J., Parrado-Hernández, E., & Ambroladze, A. (2010). Data dependent priors in PAC-Bayes bounds. In Proceedings of COMPSTAT 2010 - 19th International Conference on Computational Statistics, Keynote, Invited and Contributed Papers (pp. 231–240). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-7908-2604-3_21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free