Tensor completion via a multi-linear low-n-rank factorization model

41Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

The tensor completion problem is to recover a low-n-rank tensor from a subset of its entries. The main solution strategy has been based on the extensions of trace norm for the minimization of tensor rank via convex optimization. This strategy bears the computational cost required by the singular value decomposition (SVD) which becomes increasingly expensive as the size of the underlying tensor increase. In order to reduce the computational cost, we propose a multi-linear low-n-rank factorization model and apply the nonlinear Gauss-Seidal method that only requires solving a linear least squares problem per iteration to solve this model. Numerical results show that the proposed algorithm can reliably solve a wide range of problems at least several times faster than the trace norm minimization algorithm. © 2014 The Authors.

Cite

CITATION STYLE

APA

Tan, H., Cheng, B., Wang, W., Zhang, Y. J., & Ran, B. (2014). Tensor completion via a multi-linear low-n-rank factorization model. Neurocomputing, 133, 161–169. https://doi.org/10.1016/j.neucom.2013.11.020

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free