Evaluating compositionality of sentence representation models

4Citations
Citations of this article
73Readers
Mendeley users who have this article in their library.

Abstract

We evaluate the compositionality of general-purpose sentence encoders by proposing two metrics to quantify compositional understanding capability of sentence encoders. We introduce a novel metric, Polarity Sensitivity Scoring (PSS), which utilizes sentiment perturbations as a proxy for measuring compositionality. We then compare results from PSS with those obtained via our proposed extension of a metric called Tree Reconstruction Error (TRE) (Andreas, 2019) where compositionality is evaluated by measuring how well a true representation-producing model can be approximated by a model that explicitly combines representations of its primitives.

Cite

CITATION STYLE

APA

Bhathena, H., Willis, A., & Dass, N. (2020). Evaluating compositionality of sentence representation models. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 185–193). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.repl4nlp-1.22

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free