CoTEVer: Chain of Thought Prompting Annotation Toolkit for Explanation Verification

3Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.

Abstract

Chain-of-thought (CoT) prompting enables large language models (LLMs) to solve complex reasoning tasks by generating an explanation before the final prediction. Despite it’s promising ability, a critical downside of CoT prompting is that the performance is greatly affected by the factuality of the generated explanation. To improve the correctness of the explanations, fine-tuning language models with explanation data is needed. However, there exists only a few datasets that can be used for such approaches, and no data collection tool for building them. Thus, we introduce CoTEVer, a tool-kit for annotating the factual correctness of generated explanations and collecting revision data of wrong explanations. Furthermore, we suggest several use cases where the data collected with CoTEVer can be utilized for enhancing the faithfulness of explanations. Our toolkit is publicly available at https://github.com/SeungoneKim/CoTEVer.

Cite

CITATION STYLE

APA

Kim, S., Joo, S., Jang, Y., Chae, H., & Yeo, J. (2023). CoTEVer: Chain of Thought Prompting Annotation Toolkit for Explanation Verification. In EACL 2023 - 17th Conference of the European Chapter of the Association for Computational Linguistics, Proceedings of System Demonstrations (pp. 195–208). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.eacl-demo.23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free