Chain of Thought Prompting Elicits Knowledge Augmentation

33Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The knowledge-augmented deep learning paradigm refers to a paradigm in which domain knowledge is identified and integrated into deep models. Conventional methods typically employ task-specific approaches to gather external knowledge from various sources. In contrast, large language models are extensively pre-trained and can serve as a comprehensive source of external knowledge. In this paper, we propose CoT-KA, a Chain-of-Thought-based method that augments knowledge for deep learning. CoT-KA avoids the need for additional knowledge retrieval or knowledge reasoning models, as required in conventional augmentation methods. Our results demonstrate that CoT-KA outperforms both pure CoT-based methods and the non-augmented method across the majority of eleven publicly available benchmarks for various reasoning tasks.

Cite

CITATION STYLE

APA

Wu, D., Zhang, J., & Huang, X. (2023). Chain of Thought Prompting Elicits Knowledge Augmentation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 6519–6534). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.408

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free