PROTEST-ER: Retraining BERT for Protest Event Extraction

9Citations
Citations of this article
53Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We analyze the effect of further pre-training BERT with different domain specific data as an unsupervised domain adaptation strategy for event extraction. Portability of event extraction models is particularly challenging, with large performance drops affecting data on the same text genres (e.g., news). We present PROTEST-ER, a retrained BERT model for protest event extraction. PROTEST-ER outperforms a corresponding generic BERT on outof- domain data of 8.1 points. Our best performing models reach 51.91-46.39 F1 across both domains.

Cite

CITATION STYLE

APA

Caselli, T., Mutlu, O., Basile, A., & Urriyetôglu, A. H. (2021). PROTEST-ER: Retraining BERT for Protest Event Extraction. In 4th Workshop on Challenges and Applications of Automated Extraction of Socio-Political Events from Text, CASE 2021 - Proceedings (pp. 12–19). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.case-1.4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free