Abstract
This paper describes our system for SemEval-2021 Task 5 on Toxic Spans Detection. We developed ensemble models using BERT-based neural architectures and post-processing to combine tokens into spans. We evaluated several pre-trained language models using various ensemble techniques for toxic span identification and achieved sizable improvements over our baseline fine-tuned BERT models. Finally, our system obtained a F1-score of 67.55% on test data.
Cite
CITATION STYLE
Kotyushev, M., Glazkova, A., & Morozov, D. (2021). MIPT-NSU-UTMN at SemEval-2021 Task 5: Ensembling Learning with Pre-trained Language Models for Toxic Spans Detection. In SemEval 2021 - 15th International Workshop on Semantic Evaluation, Proceedings of the Workshop (pp. 913–918). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.semeval-1.124
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.