A leave-one-out cross validation bound for kernel methods with applications in learning

15Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we prove a general leave-one-out style crossvalidation bound for Kernel methods. We apply this bound to some classification and regression problems, and compare the results with previously known bounds. One aspect of our analysis is that the derived expected generalization bounds reflect both approximation (bias) and learning (variance) properties of the underlying kernel methods. We are thus able to demonstrate the universality of certain learning formulations.

Cite

CITATION STYLE

APA

Zhang, T. (2001). A leave-one-out cross validation bound for kernel methods with applications in learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2111, pp. 427–443). Springer Verlag. https://doi.org/10.1007/3-540-44581-1_28

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free