Divide-and-conquer learning with Nyström: Optimal rate and algorithm

14Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

Kernel Regularized Least Squares (KRLS) is a fundamental learner in machine learning. However, due to the high time and space requirements, it has no capability to large scale scenarios. Therefore, we propose DC-NY, a novel algorithm that combines divide-and-conquer method, Nyström, conjugate gradient, and preconditioning to scale up KRLS, has the same accuracy of exact KRLS and the minimum time and space complexity compared to the state-of-the-art approximate KRLS estimates. We present a theoretical analysis of DC-NY, including a novel error decomposition with the optimal statistical accuracy guarantees. Extensive experimental results on several real-world large-scale datasets containing up to 1M data points show that DC-NY significantly outperforms the state-of-the-art approximate KRLS estimates.

Cite

CITATION STYLE

APA

Yin, R., Liu, Y., Lu, L., Wang, W., & Meng, D. (2020). Divide-and-conquer learning with Nyström: Optimal rate and algorithm. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 6696–6703). AAAI press. https://doi.org/10.1609/aaai.v34i04.6147

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free