Investigating Math Word Problems using Pretrained Multilingual Language Models

6Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we revisit math word problems (MWPs) from the cross-lingual and multilingual perspective. We construct our MWP solvers over pretrained multilingual language models using the sequence-to-sequence model with copy mechanism. We compare how the MWP solvers perform in cross-lingual and multilingual scenarios. To facilitate the comparison of cross-lingual performance, we first adapt the large-scale English dataset MathQA as a counterpart of the Chinese dataset Math23K. Then we extend several English datasets to bilingual datasets through machine translation plus human annotation. Our experiments show that the MWP solvers may not be transferred to a different language even if the target expressions share the same numerical constants and operator set. However, it can be better generalized if problem types exist on both source language and target language.

Cite

CITATION STYLE

APA

Tan, M., Wang, L., Jiang, L., & Jiang, J. (2022). Investigating Math Word Problems using Pretrained Multilingual Language Models. In MathNLP 2022 - 1st Workshop on Mathematical Natural Language Processing, Proceedings of the Workshop (pp. 7–16). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.mathnlp-1.2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free