Leveraging Similar Users for Personalized Language Modeling with Limited Data

20Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.

Abstract

Personalized language models are designed and trained to capture language patterns specific to individual users. This makes them more accurate at predicting what a user will write. However, when a new user joins a platform and not enough text is available, it is harder to build effective personalized language models. We propose a solution for this problem, using a model trained on users that are similar to a new user. In this paper, we explore strategies for finding the similarity between new users and existing ones and methods for using the data from existing users who are a good match. We further explore the trade-off between available data for new users and how well their language can be modeled.

Cite

CITATION STYLE

APA

Welch, C., Gu, C., Kummerfeld, J. K., Pérez-Rosas, V., & Mihalcea, R. (2022). Leveraging Similar Users for Personalized Language Modeling with Limited Data. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 1742–1752). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.122

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free