AdaLip: An Adaptive Learning Rate Method per Layer for Stochastic Optimization

5Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Various works have been published around the optimization of Neural Networks that emphasize the significance of the learning rate. In this study we analyze the need for a different treatment for each layer and how this affects training. We propose a novel optimization technique, called AdaLip, that utilizes an estimation of the Lipschitz constant of the gradients in order to construct an adaptive learning rate per layer that can work on top of already existing optimizers, like SGD or Adam. A detailed experimental framework was used to prove the usefulness of the optimizer on three benchmark datasets. It showed that AdaLip improves the training performance and the convergence speed, but also made the training process more robust to the selection of the initial global learning rate.

Cite

CITATION STYLE

APA

Ioannou, G., Tagaris, T., & Stafylopatis, A. (2023). AdaLip: An Adaptive Learning Rate Method per Layer for Stochastic Optimization. Neural Processing Letters, 55(5), 6311–6338. https://doi.org/10.1007/s11063-022-11140-w

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free