Adapting Loss Functions to Learning Progress Improves Accuracy of Classification in Neural Networks

1Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Power error loss (PEL) has recently been suggested as a more efficient generalization of binary or categorical cross entropy (BCE/CCE). However, as PEL requires to adapt the exponent q of a power function to training data and learning progress, it has been argued that the observed improvements may be due to implicitly optimizing learning rate. Here we invalidate this argument by optimizing learning rate in each training step. We find that PEL clearly remains superior over BCE/CCE if q is properly decreased during learning. This proves that the dominant mechanism of PEL is better adapting to output error distributions, rather than implicitly manipulating learning rate.

Cite

CITATION STYLE

APA

Knoblauch, A. (2022). Adapting Loss Functions to Learning Progress Improves Accuracy of Classification in Neural Networks. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13515 LNAI, pp. 272–282). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-16564-1_26

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free