There are a growing number of table pretraining methods proposed for reasoning over tabular data (e.g., question answering, fact checking, and faithful text generation). However, most existing methods are benchmarked solely on a limited number of datasets, varying in configuration, which leads to a lack of unified, standardized, fair, and comprehensive comparison between methods. This paper presents OPENRT, the first open-source framework for reasoning over tabular data, to reproduce existing table pre-Training models for performance comparison and develop new models quickly. We implemented and compared six table pre-Training models on four question answering, one fact checking, and one faithful text generation datasets. Moreover, to enable the community to easily construct new table reasoning datasets, we developed TARAT, an annotation tool which supports multi-person collaborative annotations for various kinds of table reasoning tasks. The researchers are able to deploy the newly-constructed dataset to OPENRT and compare the performances of different baseline systems. The library OPENRT, along with the annotation tool TARAT, is publicly available at https://github.com/ yilunzhao/OpenRT.
CITATION STYLE
Zhao, Y., Mi, B., Qi, Z., Nan, L., Guo, M., Cohan, A., & Radev, D. (2023). Openrt: An open-source framework for reasoning over tables. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 3, pp. 336–347). Association for Computational Linguistics (ACL).
Mendeley helps you to discover research relevant for your work.