GeoQA: A Geometric Question Answering Benchmark Towards Multimodal Numerical Reasoning
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
Automatic math problem solving has recently attracted increasing attention as a longstanding AI benchmark. In this paper, we focus on solving geometric problems, which requires a comprehensive understanding of textual descriptions, visual diagrams, and theorem knowledge. However, the existing methods were highly dependent on handcraft rules and were merely evaluated on small-scale datasets. Therefore, we propose a Geometric Question Answering dataset GeoQA, containing 5,010 geometric problems with corresponding annotated programs, which illustrate the solving process of the given problems. Compared with another publicly available dataset GeoS, GeoQA is 25 times larger, in which the program annotations can provide a practical testbed for future research on explicit and explainable numerical reasoning. Moreover, we introduce a Neural Geometric Solver (NGS) to address geometric problems by comprehensively parsing multimodal information and generating interpretable programs. We further add multiple self-supervised auxiliary tasks on NGS to enhance cross-modal semantic representation. Extensive experiments on GeoQA validate the effectiveness of our proposed NGS and auxiliary tasks. However, the results are still significantly lower than human performance, which leaves large room for future research. Our benchmark and code are released at https://github.com/chenjudge/GeoQA.
Annotated program; Geometric problems; Hand-craft rules; Math problem solving; Multi-modal; Numerical reasoning; Program annotation; Question Answering; Small scale; Textual description
J. Chen et al., “GeoQA: A geometric question answering benchmark towards multimodal numerical reasoning,” Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 , pp. 513–523, 2021, doi: 10.18653/V1/2021.FINDINGS-ACL.46.