Learning based on kernel machines is widely known as a powerful tool for various fields of information science such as pattern recognition and regression estimation. An appropriate model selection is required in order to obtain desirable learning results. In our previous work, we discussed a class of kernels forming a nested class of reproducing kernel Hilbert spaces with an invariant metric and proved that the kernel corresponding to the smallest reproducing kernel Hilbert space, including an unknown true function, gives the best model. In this paper, we relax the invariant metric condition and show that a similar result is obtained when a subspace with an invariant metric exists. © 2012 Springer-Verlag Berlin Heidelberg.
CITATION STYLE
Tanaka, A., Takigawa, I., Imai, H., & Kudo, M. (2012). Extended analyses for an optimal kernel in a class of kernels with an invariant metric. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7626 LNCS, pp. 345–353). https://doi.org/10.1007/978-3-642-34166-3_38
Mendeley helps you to discover research relevant for your work.