TimeLMs: Diachronic Language Models from Twitter

163Citations
Citations of this article
171Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Despite its importance, the time variable has been largely neglected in the NLP and language model literature. In this paper, we present TimeLMs, a set of language models specialized on diachronic Twitter data. We show that a continual learning strategy contributes to enhancing Twitter-based language models’ capacity to deal with future and out-of-distribution tweets, while making them competitive with standardized and more monolithic benchmarks. We also perform a number of qualitative analyses showing how they cope with trends and peaks in activity involving specific named entities or concept drift. TimeLMs is available at https://github.com/cardiffnlp/timelms.

Cite

CITATION STYLE

APA

Loureiro, D., Barbieri, F., Neves, L., Anke, L. E., & Camacho-Collados, J. (2022). TimeLMs: Diachronic Language Models from Twitter. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 251–260). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-demo.25

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free