Get an A in Math: Progressive Rectification Prompting

3Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Chain-of-Thought (CoT) prompting methods have enabled large language models (LLMs) to generate reasoning paths and solve math word problems (MWPs). However, they are sensitive to mistakes in the paths, as any mistake can result in an incorrect answer. We propose a novel method named Progressive Rectification Prompting (PRP) to improve average accuracy on eight MWP datasets from 77.3 to 90.5. Given an initial answer from CoT, PRP iterates a verify-then-rectify process to progressively identify incorrect answers and rectify the reasoning paths. With the most likely correct answer, the LLM predicts a masked numerical value in the question; if the prediction does not match the masked value, the answer is likely incorrect. Then the LLM is prompted to regenerate the reasoning path hinted with a set of incorrect answers to prevent itself from repeating previous mistakes. PRP achieves the best performance compared against the CoT methods. Our implementation is made publicly available at https://wzy6642.github.io/prp.github.io/.

Cite

CITATION STYLE

APA

Wu, Z., Jiang, M., & Shen, C. (2024). Get an A in Math: Progressive Rectification Prompting. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 38, pp. 19288–19296). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v38i17.29898

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free