Masked Latent Semantic Modeling: an Efficient Pre-training Alternative to Masked Language Modeling

2Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we propose an alternative to the classic masked language modeling (MLM) pre-training paradigm, where we modify the objective from the reconstruction of the exact identity of randomly selected masked subwords to the prediction of their latent semantic properties. We coin the proposed pre-training technique masked latent semantic modeling (MLSM for short). In order to make the contextualized determination of the latent semantic properties of the masked subwords possible, we rely on an unsupervised technique using sparse coding. Our experimental results reveal that the fine-tuned performance of those models that we pre-trained via MLSM is consistently and significantly better compared to the use of vanilla MLM pre-training and other strong baselines.

Cite

CITATION STYLE

APA

Berend, G. (2023). Masked Latent Semantic Modeling: an Efficient Pre-training Alternative to Masked Language Modeling. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 13949–13962). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.876

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free