Feature extraction using low-rank approximations of the kernel matrix

5Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this work we use kernel subspace techniques to perform feature extraction. The projections of the data onto the coordinates of the high-dimensional space created by the kernel function are called features. The basis vectors to project the data depend on the eigendecomposition of the kernel matrix which might become very high-dimensional in case of a large training set. Nevertheless only the largest eigenvalues and corresponding eigenvectors are used to extract relevant features. In this work, we present low-rank approximations to the kernel matrix based on the Nyström method. Numerical simulations will then be used to demonstrate the Nyström extension method applied to feature extraction and classification. The performance of the presented methods is demonstrated using the USPS data set. © 2008 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Teixeira, A. R., Tomé, A. M., & Lang, E. W. (2008). Feature extraction using low-rank approximations of the kernel matrix. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5112 LNCS, pp. 404–412). https://doi.org/10.1007/978-3-540-69812-8_40

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free