Team Mxgra at SemEval-2020 Task 4: common sense making with next token prediction

0Citations
Citations of this article
65Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we explore solutions to a common sense making task in which a model must discern which of two sentences is against common sense. We used a pre-trained language model which we used to calculate perplexity scores for input to discern which sentence contained an unlikely sequence of tokens. Other approaches we tested were word vector distances, which were used to find semantic outliers within a sentence, and siamese network. By using the pre-trained language model to calculate perplexity scores based on the sequence of tokens in input sentences, we achieved an accuracy of 75 percent.

Cite

CITATION STYLE

APA

Collins, K., Grathwohl, M., & Ahmed, H. (2020). Team Mxgra at SemEval-2020 Task 4: common sense making with next token prediction. In 14th International Workshops on Semantic Evaluation, SemEval 2020 - co-located 28th International Conference on Computational Linguistics, COLING 2020, Proceedings (pp. 569–573). International Committee for Computational Linguistics. https://doi.org/10.18653/v1/2020.semeval-1.71

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free