Abstract
In this paper, we describe our system submitted for SemEval 2020 Task 9, Sentiment Analysis for Code-Mixed Social Media Text alongside other experiments. Our best performing system is a Transfer Learning-based model that fine-tunes XLM-RoBERTa, a transformer-based multilingual masked language model, on monolingual English and Spanish data and Spanish-English code-mixed data. Our system outperforms the official task baseline by achieving a 70.1% average F1-Score on the official leaderboard using the test set. For later submissions, our system manages to achieve a 75.9% average F1-Score on the test set using CodaLab username “ahmed0sultan”.
Cite
CITATION STYLE
Sultan, A., Salim, M., Gaber, A., & Hosary, I. E. (2020). WESSA at SemEval-2020 Task 9: Code-Mixed Sentiment Analysis using Transformers. In 14th International Workshops on Semantic Evaluation, SemEval 2020 - co-located 28th International Conference on Computational Linguistics, COLING 2020, Proceedings (pp. 1342–1347). International Committee for Computational Linguistics. https://doi.org/10.18653/v1/2020.semeval-1.181
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.