An alternating proximal splitting method with global convergence for nonconvex structured sparsity optimization

13Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.

Abstract

In many learning tasks with structural properties, structured sparse modeling usually leads to better interpretability and higher generalization performance. While great efforts have focused on the convex regularization, recent studies show that nonconvex regularizers can outperform their convex counterparts in many situations. However, the resulting nonconvex optimization problems are still challenging, especially for the structured sparsity-inducing regularizers. In this paper, we propose a splitting method for solving nonconvex structured sparsity optimization problems. The proposed method alternates between a gradient step and an easily solvable proximal step, and thus enjoys low per-iteration computational complexity. We prove that the whole sequence generated by the proposed method converges to a critical point with at least sublinear convergence rate, relying on the Kurdyka-Łojasiewicz inequality. Experiments on both simulated and real-world data sets demonstrate the efficiency and efficacy of the proposed method.

Cite

CITATION STYLE

APA

Zhang, S., Qian, H., & Gong, X. (2016). An alternating proximal splitting method with global convergence for nonconvex structured sparsity optimization. In 30th AAAI Conference on Artificial Intelligence, AAAI 2016 (pp. 2330–2336). AAAI press. https://doi.org/10.1609/aaai.v30i1.10253

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free