Implicit discourse relation classification: We need to talk about evaluation

43Citations
Citations of this article
115Readers
Mendeley users who have this article in their library.

Abstract

Implicit relation classification on Penn Discourse TreeBank (PDTB) 2.0 is a common benchmark task for evaluating the understanding of discourse relations. However, the lack of consistency in preprocessing and evaluation poses challenges to fair comparison of results in the literature. In this work, we highlight these inconsistencies and propose an improved evaluation protocol. Paired with this protocol, we report strong baseline results from pretrained sentence encoders, which set the new state-of-the-art for PDTB 2.0. Furthermore, this work is the first to explore fine-grained relation classification on PDTB 3.0. We expect our work to serve as a point of comparison for future work, and also as an initiative to discuss models of larger context and possible data augmentations for downstream transferability.

Cite

CITATION STYLE

APA

Kim, N., Feng, S., Gunasekara, C., & Lastras, L. A. (2020). Implicit discourse relation classification: We need to talk about evaluation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 5404–5414). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.480

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free