Abstract
Answering questions in narratives about why events happened often requires commonsense knowledge external to the text. What aspects of this knowledge are available in large language models? What aspects can be made accessible via external commonsense resources? We study these questions in the context of answering questions in the TELLMEWHY dataset using COMET as a source of relevant commonsense relations. We analyze the effects of model size (T5 variants and GPT-3) along with methods of injecting knowledge (COMET) into these models. Results show that the largest models, as expected, yield substantial improvements over base models and injecting external knowledge helps models of all sizes. We also find that the format in which knowledge is provided is critical, and that smaller models benefit more from larger amounts of knowledge. Finally, we develop an ontology of knowledge types and analyze the relative coverage of the models across these categories.
Cite
CITATION STYLE
Lal, Y. K., Tandon, N., Aggarwal, T., Liu, H., Chambers, N., Mooney, R., & Balasubramanian, N. (2022). Using Commonsense Knowledge to Answer Why-Questions. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 1204–1219). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.79
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.