Stochastic gradient methods for unconstrained optimization

3Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

This papers presents an overview of gradient basedmethods for minimization of noisy functions. It is assumed that the objective functions is either given with error terms of stochastic nature or given as the mathematical expectation. Such problems arise in the context of simulation based optimization. The focus of this presentation is on the gradient based Stochastic Approximation and Sample Average Approximation methods. The concept of stochastic gradient approximation of the true gradient can be successfully extended to deterministic problems. Methods of this kind are presented for the data fitting and machine learning problems.

Cite

CITATION STYLE

APA

Krejić, N., & Jerinkić, N. K. (2014). Stochastic gradient methods for unconstrained optimization. Pesquisa Operacional, 34(3), 373–393. https://doi.org/10.1590/0101-7438.2014.034.03.0373

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free