Towards Understanding Large-Scale Discourse Structures in Pre-Trained and Fine-Tuned Language Models

10Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.

Abstract

With a growing number of BERTology works analyzing different components of pre-trained language models, we extend this line of research through an in-depth analysis of discourse information in pre-trained and fine-tuned language models. We move beyond prior work along three dimensions: First, we describe a novel approach to infer discourse structures from arbitrarily long documents. Second, we propose a new type of analysis to explore where and how accurately intrinsic discourse is captured in the BERT and BART models. Finally, we assess how similar the generated structures are to a variety of baselines as well as their distributions within and between models.

Cite

CITATION STYLE

APA

Huber, P., & Carenini, G. (2022). Towards Understanding Large-Scale Discourse Structures in Pre-Trained and Fine-Tuned Language Models. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 2376–2394). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.170

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free