Measuring and Improving BERT’s Mathematical Abilities by Predicting the Order of Reasoning.

Piotr Piękos, Mateusz Malinowski, Henryk Michalewski


Abstract
Imagine you are in a supermarket. You have two bananas in your basket and want to buy four apples. How many fruits do you have in total? This seemingly straightforward question can be challenging for data-driven language models, even if trained at scale. However, we would expect such generic language models to possess some mathematical abilities in addition to typical linguistic competence. Towards this goal, we investigate if a commonly used language model, BERT, possesses such mathematical abilities and, if so, to what degree. For that, we fine-tune BERT on a popular dataset for word math problems, AQuA-RAT, and conduct several tests to understand learned representations better. Since we teach models trained on natural language to do formal mathematics, we hypothesize that such models would benefit from training on semi-formal steps that explain how math results are derived. To better accommodate such training, we also propose new pretext tasks for learning mathematical rules. We call them (Neighbor) Reasoning Order Prediction (ROP or NROP). With this new model, we achieve significantly better outcomes than data-driven baselines and even on-par with more tailored models.
Anthology ID:
2021.acl-short.49
Volume:
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)
Month:
August
Year:
2021
Address:
Online
Venues:
ACL | IJCNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
383–394
Language:
URL:
https://aclanthology.org/2021.acl-short.49
DOI:
10.18653/v1/2021.acl-short.49
Bibkey:
Copy Citation:
PDF:
https://aclanthology.org/2021.acl-short.49.pdf
Optional supplementary material:
 2021.acl-short.49.OptionalSupplementaryMaterial.zip
Data
AQUA-RAT