Unsupervised Methods for the Study of Transformer Embeddings

2Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Over the last decade neural word embeddings have become a cornerstone of many important text mining applications such as text classification, sentiment analysis, named entity recognition, question answering systems, etc. Particularly, Transformer-based contextual word embeddings have gained much attention with several works trying to understanding how such models work, through the use of supervised probing tasks, and usually emphasizing on BERT. In this paper, we propose a fully unsupervised manner to analyze Transformer-based embedding models in their bare state with no fine-tuning. We more precisely focus on characterizing and identifying groups of Transformer layers across 6 different Transformer models.

Cite

CITATION STYLE

APA

Ait Saada, M., Role, F., & Nadif, M. (2021). Unsupervised Methods for the Study of Transformer Embeddings. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12695 LNCS, pp. 287–300). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-74251-5_23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free