Generalized dantzig selector: Application to the K-support norm

ISSN: 10495258
28Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.

Abstract

We propose a Generalized Dantzig Selector (GDS) for linear models, in which any norm encoding the parameter structure can be leveraged for estimation. We investigate both computational and statistical aspects of the GDS. Based on conjugate proximal operator, a flexible inexact ADMM framework is designed for solving GDS. Thereafter, non-asymptotic high-probability bounds are established on the estimation error, which rely on Gaussian widths of the unit norm ball and the error set. Further, we consider a non-trivial example of the GDS using k-support norm. We derive an efficient method to compute the proximal operator for k-support norm since existing methods are inapplicable in this setting. For statistical analysis, we provide upper bounds for the Gaussian widths needed in the GDS analysis, yielding the first statistical recovery guarantee for estimation with the k-support norm. The experimental results confirm our theoretical analysis.

Cite

CITATION STYLE

APA

Chatterjee, S., Chen, S., & Banerjee, A. (2014). Generalized dantzig selector: Application to the K-support norm. In Advances in Neural Information Processing Systems (Vol. 3, pp. 1934–1942). Neural information processing systems foundation.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free