Does Head Label Help for Long-Tailed Multi-Label Text Classification

58Citations
Citations of this article
54Readers
Mendeley users who have this article in their library.

Abstract

Multi-label text classification (MLTC) aims to annotate documents with the most relevant labels from a number of candidate labels. In real applications, the distribution of label frequency often exhibits a long tail, i.e., a few labels are associated with a large number of documents (a.k.a. head labels), while a large fraction of labels are associated with a small number of documents (a.k.a. tail labels). To address the challenge of insufficient training data on tail label classification, we propose a Head-to-Tail Network (HTTN) to transfer the meta-knowledge from the data-rich head labels to data-poor tail labels. The meta-knowledge is the mapping from fewshot network parameters to many-shot network parameters, which aims to promote the generalizability of tail classifiers. Extensive experimental results on three benchmark datasets demonstrate that HTTN consistently outperforms the stateof- the-art methods. The code and hyper-parameter settings are released for reproducibility.

Cite

CITATION STYLE

APA

Xiao, L., Zhang, X., Jing, L., Huang, C., & Song, M. (2021). Does Head Label Help for Long-Tailed Multi-Label Text Classification. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 16, pp. 14103–14111). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i16.17660

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free