Why Do Document-Level Polarity Classifiers Fail?

4Citations
Citations of this article
51Readers
Mendeley users who have this article in their library.

Abstract

Machine learning solutions are often criticized for the lack of explanation of their successes and failures. Understanding which instances are misclassified and why is essential to improve the learning process. This work helps to fill this gap by proposing a methodology to characterize, quantify and measure the impact of hard instances in the task of polarity classification of movie reviews. We characterize such instances into two categories: neutrality, where the text does not convey a clear polarity, and discrepancy, where the polarity of the text is the opposite of its true rating. We quantify the number of hard instances in polarity classification of movie reviews and provide empirical evidence about the need to pay attention to such problematic instances, as they are much harder to classify, for both machine and human classifiers. To the best of our knowledge, this is the first systematic analysis of the impact of hard instances in polarity detection from well-formed textual reviews.

Cite

CITATION STYLE

APA

Martins, K. S., Vaz-De-Melo, P. O. S., & Santos, R. L. T. (2021). Why Do Document-Level Polarity Classifiers Fail? In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1782–1794). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.143

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free