Approximating the Gradient of Cross-Entropy Loss Function

68Citations
Citations of this article
52Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

A loss function has two crucial roles in training a conventional discriminant deep neural network (DNN): (i) it measures the goodness of classification and (ii) generates the gradients that drive the training of the network. In this paper, we approximate the gradients of cross-entropy loss which is the most often used loss function in the classification DNNs. The proposed approximations are noise-free, which means they depend only on the labels of the training set. They have a fixed length to avoid the vanishing gradient problem of the cross-entropy loss. By skipping the forward pass, the computational complexities of the proposed approximations are reduced to mathcal {O}(n) where n is the batch size. Two claims are established based on the experiments of training DNNs using the proposed approximations: (i) It is possible to train a discriminant network without explicitly defining a loss function. (ii) The success of training does not imply the convergence of network parameters to fixed values. The experiments show that the proposed gradient approximations achieve comparable classification accuracy to the conventional loss functions and can accelerate the training process on multiple datasets.

Cite

CITATION STYLE

APA

Li, L., Doroslovacki, M., & Loew, M. H. (2020). Approximating the Gradient of Cross-Entropy Loss Function. IEEE Access, 8, 111626–111635. https://doi.org/10.1109/ACCESS.2020.3001531

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free