Abstract
We propose a variant of Convolutional Neural Network (CNN) models, the Attention CNN (ACNN); for large-scale categorization of millions of Japanese items into thirty-five product categories. Compared to a state-of-the-art Gradient Boosted Tree (GBT) classifier, the proposed model reduces training time from three weeks to three days while maintaining more than 96% accuracy. Additionally, our proposed model characterizes products by imputing attentive focus on word tokens in a language agnostic way. The attention words have been observed to be semantically highly correlated with the predicted categories and give us a choice of automatic feature extraction for downstream processing.
Cite
CITATION STYLE
Xia, Y., Levine, A., Das, P., Di Fabbrizio, G., Shinzato, K., & Datta, A. (2017). Large-scale categorization of Japanese product titles using neural attention models. In 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017 - Proceedings of Conference (Vol. 2, pp. 663–668). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/e17-2105
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.