Fast optimization methods for L1 regularization: A comparative study and two new approaches

217Citations
Citations of this article
268Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

L1 regularization is effective for feature selection, but the resulting optimization is challenging due to the non-differentiability of the 1-norm. In this paper we compare state-of-the-art optimization techniques to solve this problem across several loss functions. Furthermore, we propose two new techniques. The first is based on a smooth (differentiable) convex approximation for the L1 regularizer that does not depend on any assumptions about the loss function used. The other technique is a new strategy that addresses the non-differentiability of the L1-regularizer by casting the problem as a constrained optimization problem that is then solved using a specialized gradient projection method. Extensive comparisons show that our newly proposed approaches consistently rank among the best in terms of convergence speed and efficiency by measuring the number of function evaluations required. © Springer-Verlag Berlin Heidelberg 2007.

Cite

CITATION STYLE

APA

Schmidt, M., Fung, G., & Rosales, R. (2007). Fast optimization methods for L1 regularization: A comparative study and two new approaches. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4701 LNAI, pp. 286–297). Springer Verlag. https://doi.org/10.1007/978-3-540-74958-5_28

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free