Learning Multi-Step Reasoning by Solving Arithmetic Tasks

2 Jun 2023  ·  Tianduo Wang, Wei Lu ·

Mathematical reasoning is regarded as a necessary ability for Language Models (LMs). Recent works demonstrate large LMs' impressive performance in solving math problems. The success is attributed to their Chain-of-Thought (CoT) reasoning abilities, i.e., the ability to decompose complex questions into step-by-step reasoning chains, but such ability seems only to emerge from models with abundant parameters. This work investigates how to incorporate relatively small LMs with the capabilities of multi-step reasoning. We propose to inject such abilities by continually pre-training LMs on a synthetic dataset MsAT which is composed of Multi-step Arithmetic Tasks. Our experiments on four math word problem datasets show the effectiveness of the proposed method in enhancing LMs' math reasoning abilities.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Math Word Problem Solving MAWPS MsAT-DeductReasoner Accuracy (%) 94.3 # 2
Math Word Problem Solving SVAMP MsAT-DeductReasoner Execution Accuracy 48.9 # 13

Methods


No methods listed for this paper. Add relevant methods here