A Geometric Integration Approach to Nonsmooth, Nonconvex Optimisation

9Citations
Citations of this article
28Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The optimisation of nonsmooth, nonconvex functions without access to gradients is a particularly challenging problem that is frequently encountered, for example in model parameter optimisation problems. Bilevel optimisation of parameters is a standard setting in areas such as variational regularisation problems and supervised machine learning. We present efficient and robust derivative-free methods called randomised Itoh–Abe methods. These are generalisations of the Itoh–Abe discrete gradient method, a well-known scheme from geometric integration, which has previously only been considered in the smooth setting. We demonstrate that the method and its favourable energy dissipation properties are well defined in the nonsmooth setting. Furthermore, we prove that whenever the objective function is locally Lipschitz continuous, the iterates almost surely converge to a connected set of Clarke stationary points. We present an implementation of the methods, and apply it to various test problems. The numerical results indicate that the randomised Itoh–Abe methods can be superior to state-of-the-art derivative-free optimisation methods in solving nonsmooth problems while still remaining competitive in terms of efficiency.

Cite

CITATION STYLE

APA

Riis, E. S., Ehrhardt, M. J., Quispel, G. R. W., & Schönlieb, C. B. (2022). A Geometric Integration Approach to Nonsmooth, Nonconvex Optimisation. Foundations of Computational Mathematics, 22(5), 1351–1394. https://doi.org/10.1007/s10208-020-09489-2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free