TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Teaching-Inspired Integrated Prompting Framework: A Novel ...

Teaching-Inspired Integrated Prompting Framework: A Novel Approach for Enhancing Reasoning in Large Language Models

Wenting Tan, Dongxiao Chen, Jieting Xue, ZiHao Wang, Taijie Chen

2024-10-10Mathematical ReasoningMathMath Word Problem SolvingArithmetic Reasoning
PaperPDFCode(official)

Abstract

Large Language Models (LLMs) exhibit impressive performance across various domains but still struggle with arithmetic reasoning tasks. Recent work shows the effectiveness of prompt design methods in enhancing reasoning capabilities. However, these approaches overlook crucial requirements for prior knowledge of specific concepts, theorems, and tricks to tackle most arithmetic reasoning problems successfully. To address this issue, we propose a novel and effective Teaching-Inspired Integrated Framework, which emulates the instructional process of a teacher guiding students. This method equips LLMs with essential concepts, relevant theorems, and similar problems with analogous solution approaches, facilitating the enhancement of reasoning abilities. Additionally, we introduce two new Chinese datasets, MathMC and MathToF, both with detailed explanations and answers. Experiments are conducted on nine benchmarks which demonstrates that our approach improves the reasoning accuracy of LLMs. With GPT-4 and our framework, we achieve new state-of-the-art performance on four math benchmarks (AddSub, SVAMP, Math23K and AQuA) with accuracies of 98.2% (+3.3%), 93.9% (+0.2%), 94.3% (+7.2%) and 81.1% (+1.2%). Our data and code are available at https://github.com/SallyTan13/Teaching-Inspired-Prompting.

Results

TaskDatasetMetricValueModel
Question AnsweringMath23KAccuracy (5-fold)94.3GPT-4 (Teaching-Inspired)
Question AnsweringSVAMPExecution Accuracy93.9GPT-4 (Teaching-Inspired)
Math Word Problem SolvingMath23KAccuracy (5-fold)94.3GPT-4 (Teaching-Inspired)
Math Word Problem SolvingSVAMPExecution Accuracy93.9GPT-4 (Teaching-Inspired)
Mathematical Question AnsweringMath23KAccuracy (5-fold)94.3GPT-4 (Teaching-Inspired)
Mathematical Question AnsweringSVAMPExecution Accuracy93.9GPT-4 (Teaching-Inspired)
Mathematical ReasoningMath23KAccuracy (5-fold)94.3GPT-4 (Teaching-Inspired)
Mathematical ReasoningSVAMPExecution Accuracy93.9GPT-4 (Teaching-Inspired)
Arithmetic ReasoningMathToFAccuracy89.2GPT-4 (Teaching-Inspired)
Arithmetic ReasoningMathMCAccuracy92.2GPT-4 (Teaching-Inspired)
Arithmetic ReasoningGSM8KAccuracy94.8GPT-4 (Teaching-Inspired)

Related Papers

VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17A Survey of Deep Learning for Geometry Problem Solving2025-07-16Scaling Up RL: Unlocking Diverse Reasoning in LLMs via Prolonged Training2025-07-16KisMATH: Do LLMs Have Knowledge of Implicit Structures in Mathematical Reasoning?2025-07-15Temperature and Persona Shape LLM Agent Consensus With Minimal Accuracy Gains in Qualitative Coding2025-07-15Personalized Exercise Recommendation with Semantically-Grounded Knowledge Tracing2025-07-15DCR: Quantifying Data Contamination in LLMs Evaluation2025-07-15