TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MWP-BERT: Numeracy-Augmented Pre-training for Math Word Pr...

MWP-BERT: Numeracy-Augmented Pre-training for Math Word Problem Solving

Zhenwen Liang, Jipeng Zhang, Lei Wang, Wei Qin, Yunshi Lan, Jie Shao, Xiangliang Zhang

2021-07-28Findings (NAACL) 2022 7MathMath Word Problem SolvingRepresentation LearningCommon Sense ReasoningProgram SynthesisLanguage Modelling
PaperPDFCode(official)

Abstract

Math word problem (MWP) solving faces a dilemma in number representation learning. In order to avoid the number representation issue and reduce the search space of feasible solutions, existing works striving for MWP solving usually replace real numbers with symbolic placeholders to focus on logic reasoning. However, different from common symbolic reasoning tasks like program synthesis and knowledge graph reasoning, MWP solving has extra requirements in numerical reasoning. In other words, instead of the number value itself, it is the reusable numerical property that matters more in numerical reasoning. Therefore, we argue that injecting numerical properties into symbolic placeholders with contextualized representation learning schema can provide a way out of the dilemma in the number representation issue here. In this work, we introduce this idea to the popular pre-training language model (PLM) techniques and build MWP-BERT, an effective contextual number representation PLM. We demonstrate the effectiveness of our MWP-BERT on MWP solving and several MWP-specific understanding tasks on both English and Chinese benchmarks.

Results

TaskDatasetMetricValueModel
Question AnsweringMath23KAccuracy (5-fold)82.4MWP-BERT
Question AnsweringMath23KAccuracy (training-test)84.7MWP-BERT
Question AnsweringMathQAAnswer Accuracy76.6MWP-BERT
Math Word Problem SolvingMath23KAccuracy (5-fold)82.4MWP-BERT
Math Word Problem SolvingMath23KAccuracy (training-test)84.7MWP-BERT
Math Word Problem SolvingMathQAAnswer Accuracy76.6MWP-BERT
Mathematical Question AnsweringMath23KAccuracy (5-fold)82.4MWP-BERT
Mathematical Question AnsweringMath23KAccuracy (training-test)84.7MWP-BERT
Mathematical Question AnsweringMathQAAnswer Accuracy76.6MWP-BERT
Mathematical ReasoningMath23KAccuracy (5-fold)82.4MWP-BERT
Mathematical ReasoningMath23KAccuracy (training-test)84.7MWP-BERT
Mathematical ReasoningMathQAAnswer Accuracy76.6MWP-BERT

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17Comparing Apples to Oranges: A Dataset & Analysis of LLM Humour Understanding from Traditional Puns to Topical Jokes2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17