Subgradient methods for huge-scale optimization problems

48Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We consider a new class of huge-scale problems, the problems with sparse subgradients. The most important functions of this type are piece-wise linear. For optimization problems with uniform sparsity of corresponding linear operators, we suggest a very efficient implementation of subgradient iterations, which total cost depends logarithmically in the dimension. This technique is based on a recursive update of the results of matrix/vector products and the values of symmetric functions. It works well, for example, for matrices with few nonzero diagonals and for max-type functions. We show that the updating technique can be efficiently coupled with the simplest subgradient methods, the unconstrained minimization method by B.Polyak, and the constrained minimization scheme by N.Shor. Similar results can be obtained for a new nonsmooth random variant of a coordinate descent scheme. We present also the promising results of preliminary computational experiments. © 2013 Springer-Verlag Berlin Heidelberg and Mathematical Optimization Society.

Cite

CITATION STYLE

APA

Nesterov, Y. (2014). Subgradient methods for huge-scale optimization problems. Mathematical Programming, 146(1–2), 275–297. https://doi.org/10.1007/s10107-013-0686-4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free