A conjugate gradient method with descent properties under strong Wolfe line search

3Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The conjugate gradient (CG) method is one of the optimization methods that are often used in practical applications. The continuous and numerous studies conducted on the CG method have led to vast improvements in its convergence properties and efficiency. In this paper, a new CG method possessing the sufficient descent and global convergence properties is proposed. The efficiency of the new CG algorithm relative to the existing CG methods is evaluated by testing them all on a set of test functions using MATLAB. The tests are measured in terms of iteration numbers and CPU time under strong Wolfe line search. Overall, this new method performs efficiently and comparable to the other famous methods.

Cite

CITATION STYLE

APA

Zull, N., Aini, N., Shoid, S., Ghani, N. H. A., Mohamed, N. S., Rivaie, M., & Mamat, M. (2017). A conjugate gradient method with descent properties under strong Wolfe line search. In Journal of Physics: Conference Series (Vol. 890). Institute of Physics Publishing. https://doi.org/10.1088/1742-6596/890/1/012105

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free