Abstract
Toxic language is often present in online forums, especially when politics and other polarizing topics arise, and can lead to people becoming discouraged from joining or continuing conversations. In this paper, I use data consisting of comments with the indices of toxic text labelled to train an RNN to determine which parts of the comments make them toxic, which could aid online moderators. I compare results using both the original dataset and an augmented set, as well as GRU versus LSTM RNN models.
Cite
CITATION STYLE
Cech, M. (2021). macech at SemEval-2021 Task 5: Toxic Spans Detection. In SemEval 2021 - 15th International Workshop on Semantic Evaluation, Proceedings of the Workshop (pp. 1003–1008). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.semeval-1.137
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.