Softmax dissection: Towards understanding intra: Towards inter-class objective for embedding learning

18Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.

Abstract

The softmax loss and its variants are widely used as objectives for embedding learning applications like face recognition. However, the intra- and inter-class objectives in Softmax are entangled, therefore a well-optimized inter-class objective leads to relaxation on the intra-class objective, and vice versa. In this paper, we propose to dissect Softmax into independent intra- and inter-class objective (D-Softmax) with a clear understanding. It is straightforward to tune each part to the best state with D-Softmax as objective.Furthermore, we find the computation of the inter-class part is redundant and propose sampling-based variants of D-Softmax to reduce the computation cost. The face recognition experiments on regularscale data show D-Softmax is favorably comparable to existing losses such as SphereFace and ArcFace. Experiments on massive-scale data show the fast variants significantly accelerates the training process (such as 64×) with only a minor sacrifice in performance, outperforming existing acceleration methods of Softmax in terms of both performance and efficiency.

Cite

CITATION STYLE

APA

He, L., Wang, Z., Li, Y., & Wang, S. (2020). Softmax dissection: Towards understanding intra: Towards inter-class objective for embedding learning. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 10957–10964). AAAI press. https://doi.org/10.1609/aaai.v34i07.6729

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free