Deriving the kernel from training data

2Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper we propose a strategy for constructing datadriven kernels, automatically determined by the training examples. Basically, their associated Reproducing Kernel Hubert Spaces arise from finite sets of linearly independent functions, that can be interpreted as weak classifiers or regressors, learned from training material. When working in the Tikhonov regularization framework, the unique free parameter to be optimized is the regularizer, representing a trade-off between empirical error and smoothness of the solution. A generalization error bound based on Rademacher complexity is provided, yielding the potential for controlling overfitting. © Springer-Verlag Berlin Heidelberg 2007.

Cite

CITATION STYLE

APA

Merler, S., Jurman, G., & Furlanello, C. (2007). Deriving the kernel from training data. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4472 LNCS, pp. 32–41). Springer Verlag. https://doi.org/10.1007/978-3-540-72523-7_4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free