Invertible nonlinear dimensionality reduction via joint dictionary learning

3Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper proposes an invertible nonlinear dimensionality reduction method via jointly learning dictionaries in both the original high dimensional data space and its low dimensional representation space. We construct an appropriate cost function, which preserves inner products of data representations in the low dimensional space. We employ a conjugate gradient algorithm on smooth manifold to minimize the cost function. By numerical experiments in image processing, our proposed method provides competitive and robust performance in image compression and recovery, even on heavily corrupted data. In other words, it can also be considered as an alternative approach to compressed sensing. While our approach can outperform compressed sensing in task-driven learning problems, such as data visualization.

Cite

CITATION STYLE

APA

Wei, X., Kleinsteuber, M., & Shen, H. (2015). Invertible nonlinear dimensionality reduction via joint dictionary learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9237, pp. 279–286). Springer Verlag. https://doi.org/10.1007/978-3-319-22482-4_32

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free