In this paper, we propose a stochastic method for solving equality constrained optimization problems that utilizes predictive variance reduction. Specifically, we develop a method based on the sequential quadratic programming paradigm that employs variance reduction in the gradient approximations. Under reasonable assumptions, we prove that a measure of first-order stationarity evaluated at the iterates generated by our proposed algorithm converges to zero in expectation from arbitrary starting points, for both constant and adaptive step size strategies. Finally, we demonstrate the practical performance of our proposed algorithm on constrained binary classification problems that arise in machine learning.
CITATION STYLE
Berahas, A. S., Shi, J., Yi, Z., & Zhou, B. (2023). Accelerating stochastic sequential quadratic programming for equality constrained optimization using predictive variance reduction. Computational Optimization and Applications, 86(1), 79–116. https://doi.org/10.1007/s10589-023-00483-2
Mendeley helps you to discover research relevant for your work.