Temporal Effects on Pre-trained Models for Language Processing Tasks

37Citations
Citations of this article
44Readers
Mendeley users who have this article in their library.

Abstract

Keeping the performance of language technologies optimal as time passes is of great practical interest. We study temporal effects on model performance on downstream language tasks, establishing a nuanced terminology for such discussion and identifying factors essen-tial to conduct a robust study. We present experiments for several tasks in English where the label correctness is not dependent on time and demonstrate the importance of distinguish-ing between temporal model deterioration and temporal domain adaptation for systems using pre-trained representations. We find that, depending on the task, temporal model deterioration is not necessarily a concern. Temporal domain adaptation, however, is beneficial in all cases, with better performance for a given time period possible when the system is trained on temporally more recent data. Therefore, we also examine the efficacy of two approaches for temporal domain adaptation without human annotations on new data. Self-labeling shows consistent improvement and notably, for named entity recognition, leads to better temporal adaptation than even human annotations.

Cite

CITATION STYLE

APA

Agarwal, O., & Nenkova, A. (2022). Temporal Effects on Pre-trained Models for Language Processing Tasks. Transactions of the Association for Computational Linguistics, 10, 904–921. https://doi.org/10.1162/tacl_a_00497

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free