Local regularizer improves generalization

3Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Regularization plays an important role in generalization of deep learning. In this paper, we study the generalization power of an unbiased regularizor for training algorithms in deep learning. We focus on training methods called Locally Regularized Stochastic Gradient Descent (LRSGD). An LRSGD leverages a proximal type penalty in gradient descent steps to regularize SGD in training. We show that by carefully choosing relevant parameters, LRSGD generalizes better than SGD. Our thorough theoretical analysis is supported by experimental evidence. It advances our theoretical understanding of deep learning and provides new perspectives on designing training algorithms. The code is available at https://github.com/huiqu18/LRSGD.

Cite

CITATION STYLE

APA

Zhang, Y., Qu, H., Metaxas, D., & Chen, C. (2020). Local regularizer improves generalization. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 6861–6868). AAAI press. https://doi.org/10.1609/aaai.v34i04.6167

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free