Personalized language models are designed and trained to capture language patterns specific to individual users. This makes them more accurate at predicting what a user will write. However, when a new user joins a platform and not enough text is available, it is harder to build effective personalized language models. We propose a solution for this problem, using a model trained on users that are similar to a new user. In this paper, we explore strategies for finding the similarity between new users and existing ones and methods for using the data from existing users who are a good match. We further explore the trade-off between available data for new users and how well their language can be modeled.
CITATION STYLE
Welch, C., Gu, C., Kummerfeld, J. K., Pérez-Rosas, V., & Mihalcea, R. (2022). Leveraging Similar Users for Personalized Language Modeling with Limited Data. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 1742–1752). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.122
Mendeley helps you to discover research relevant for your work.