Similarity or deeper understanding? Analyzing the TED-Q dataset of evoked questions

0Citations
Citations of this article
61Readers
Mendeley users who have this article in their library.

Abstract

We take a close look at a recent dataset of TED-talks annotated with the questions they implicitly evoke, TED-Q (Westera et al., 2020). We test to what extent the relation between a discourse and the questions it evokes is merely one of similarity or association, as opposed to deeper semantic/pragmatic interpretation. We do so by turning the TED-Q dataset into a binary classification task, constructing an analogous task from explicit questions we extract from the BookCorpus (Zhu et al., 2015), and fitting a BERT-based classifier alongside models based on different notions of similarity. The BERT-based classifier, achieving close to human performance, outperforms all similarity-based models, suggesting that there is more to identifying true evoked questions than plain similarity.

Cite

CITATION STYLE

APA

Westera, M., Amidei, J., & Mayol, L. (2020). Similarity or deeper understanding? Analyzing the TED-Q dataset of evoked questions. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 5004–5012). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.439

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free