The softmax loss and its variants are widely used as objectives for embedding learning applications like face recognition. However, the intra- and inter-class objectives in Softmax are entangled, therefore a well-optimized inter-class objective leads to relaxation on the intra-class objective, and vice versa. In this paper, we propose to dissect Softmax into independent intra- and inter-class objective (D-Softmax) with a clear understanding. It is straightforward to tune each part to the best state with D-Softmax as objective.Furthermore, we find the computation of the inter-class part is redundant and propose sampling-based variants of D-Softmax to reduce the computation cost. The face recognition experiments on regularscale data show D-Softmax is favorably comparable to existing losses such as SphereFace and ArcFace. Experiments on massive-scale data show the fast variants significantly accelerates the training process (such as 64×) with only a minor sacrifice in performance, outperforming existing acceleration methods of Softmax in terms of both performance and efficiency.
CITATION STYLE
He, L., Wang, Z., Li, Y., & Wang, S. (2020). Softmax dissection: Towards understanding intra: Towards inter-class objective for embedding learning. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 10957–10964). AAAI press. https://doi.org/10.1609/aaai.v34i07.6729
Mendeley helps you to discover research relevant for your work.