Investigating Transformer guided Chaining for Interpretable Natural Logic Reasoning

0Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Natural logic reasoning has received increasing attention lately, with several datasets and neural models proposed, though with limited success. More recently, a new class of works have emerged adopting a Neuro-Symbolic approach, called transformer guided chaining, whereby the idea is to iteratively perform 1-step neural inferences and chain together the results to generate a multi-step reasoning trace. Several works have adapted variants of this central idea and reported significantly high accuracies compared to vanilla LLM's. In this paper, we perform a critical empirical investigation of the chaining approach on a multi-hop First-Order Logic (FOL) reasoning benchmark. In particular, we develop a reference implementation, called Chainformer, and conduct several experiments to analyze the accuracy, generalization, interpretability, and performance over FOLs. Our findings highlight key strengths and possible current limitations and suggest potential areas for future research in logic reasoning.

Cite

CITATION STYLE

APA

Kanagasabai, R., Rajamanickam, S., & Wei, S. (2023). Investigating Transformer guided Chaining for Interpretable Natural Logic Reasoning. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 9240–9253). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.588

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free