On robust regression with high-dimensional predictors

128Citations
Citations of this article
79Readers
Mendeley users who have this article in their library.

Abstract

We study regression M-estimates in the setting where p, the number of covariates, and n, the number of observations, are both large, but p ≤ n. We find an exact stochastic representation for the distribution of β = argminaβ∈ℝp ∑i=1n ρ (Yi - X′iβ) at fixed p and n under various assumptions on the objective function ρ and our statistical model. A scalar random variable whose deterministic limit rρ (κ) can be studied when p=n → κ > 0 plays a central role in this representation. We discover a nonlinear system of two deterministic equations that characterizes rρ (κ). Interestingly, the system shows that rρ (κ) depends on ρ through proximal mappings of ρ as well as various aspects of the statistical model underlying our study. Several surprising results emerge. In particular, we show that, when p=n is large enough, least squares becomes preferable to least absolute deviations for double-exponential errors.

Cite

CITATION STYLE

APA

El Karoui, N., Bean, D., Bickel, P. J., Lim, C., & Yu, B. (2013). On robust regression with high-dimensional predictors. Proceedings of the National Academy of Sciences of the United States of America, 110(36), 14557–14562. https://doi.org/10.1073/pnas.1307842110

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free