Diff-Explainer: Differentiable Convex Optimization for Explainable Multi-hop Inference

6Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

This paper presents Diff-Explainer, the first hybrid framework for explainable multi-hop inference that integrates explicit constraints with neural architectures through differentiable convex optimization. Specifically, DiffExplainer allows for the fine-tuning of neural representations within a constrained optimization framework to answer and explain multi-hop questions in natural language. To demonstrate the efficacy of the hybrid framework, we combine existing ILP-based solvers for multi-hop Question Answering (QA) with Transformer-based representations. An extensive empirical evaluation on scientific and commonsense QA tasks demonstrates that the integration of explicit constraints in a end-to-end differentiable framework can significantly improve the performance of nondifferentiable ILP solvers (8.91%–13.3%). Moreover, additional analysis reveals that Diff-Explainer is able to achieve strong performance when compared to standalone Transformers and previous multi-hop approaches while still providing structured explanations in support of its predictions.

Cite

CITATION STYLE

APA

Thayaparan, M., Valentino, M., Ferreira, D., Rozanova, J., & Freitas, A. (2022). Diff-Explainer: Differentiable Convex Optimization for Explainable Multi-hop Inference. Transactions of the Association for Computational Linguistics, 10, 1103–1119. https://doi.org/10.1162/tacl_a_00508

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free