Abstract
Recognizing the language of ambiguous texts has become a main challenge in language identification (LID). When using multilingual applications, users have their own language preferences, which can be regarded as external knowledge for LID. Nevertheless, current studies do not consider the inter-personal variations due to the lack of user annotated training data. To fill this gap, we introduce preference-aware LID and propose a novel unsupervised learning strategy. Concretely, we construct pseudo training set for each user by extracting training samples from a standard LID corpus according to his/her historical language distribution. Besides, we contribute the first user labeled LID test set called “U-LID”. Experimental results reveal that our model can incarnate user traits and significantly outperforms existing LID systems on handling ambiguous texts. Our code and benchmark have been released.
Cite
CITATION STYLE
Ren, X., Yang, B., Liu, D., Zhang, H., Lv, X., Yao, L., & Xie, J. (2022). Unsupervised Preference-Aware Language Identification. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 3847–3852). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-acl.303
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.