Using Commonsense Knowledge to Answer Why-Questions

13Citations
Citations of this article
28Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Answering questions in narratives about why events happened often requires commonsense knowledge external to the text. What aspects of this knowledge are available in large language models? What aspects can be made accessible via external commonsense resources? We study these questions in the context of answering questions in the TELLMEWHY dataset using COMET as a source of relevant commonsense relations. We analyze the effects of model size (T5 variants and GPT-3) along with methods of injecting knowledge (COMET) into these models. Results show that the largest models, as expected, yield substantial improvements over base models and injecting external knowledge helps models of all sizes. We also find that the format in which knowledge is provided is critical, and that smaller models benefit more from larger amounts of knowledge. Finally, we develop an ontology of knowledge types and analyze the relative coverage of the models across these categories.

Cite

CITATION STYLE

APA

Lal, Y. K., Tandon, N., Aggarwal, T., Liu, H., Chambers, N., Mooney, R., & Balasubramanian, N. (2022). Using Commonsense Knowledge to Answer Why-Questions. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 1204–1219). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.79

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free