TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Progressive-Hint Prompting Improves Reasoning in Large Lan...

Progressive-Hint Prompting Improves Reasoning in Large Language Models

Chuanyang Zheng, Zhengying Liu, Enze Xie, Zhenguo Li, Yu Li

2023-04-19MathMath Word Problem SolvingGSM8KArithmetic Reasoning
PaperPDFCode(official)

Abstract

The performance of Large Language Models (LLMs) in reasoning tasks depends heavily on prompt design, with Chain-of-Thought (CoT) and self-consistency being critical methods that enhance this ability. However, these methods do not fully exploit the answers generated by the LLM to guide subsequent responses. This paper proposes a new prompting method, named Progressive-Hint Prompting (PHP), that enables automatic multiple interactions between users and LLMs by using previously generated answers as hints to progressively guide toward the correct answers. PHP is orthogonal to CoT and self-consistency, making it easy to combine with state-of-the-art techniques to further improve performance. We conducted extensive and comprehensive experiments on seven benchmarks. The results show that PHP significantly improves accuracy while remaining highly efficient. For instance, with text-davinci-003, we observed a 4.2% improvement on GSM8K with greedy decoding compared to Complex CoT, and a 46.17% reduction in sample paths with self-consistency. With GPT-4 and PHP, we achieve state-of-the-art performances on SVAMP (89.1% -> 91.9%), GSM8K (92% -> 95.5%), AQuA (76.4% -> 79.9%) and MATH (50.3% -> 53.9%).

Results

TaskDatasetMetricValueModel
Question AnsweringMATHAccuracy53.9PHP (GPT-4 model)
Question AnsweringSVAMPExecution Accuracy91.9GPT-4 (PHP)
Math Word Problem SolvingMATHAccuracy53.9PHP (GPT-4 model)
Math Word Problem SolvingSVAMPExecution Accuracy91.9GPT-4 (PHP)
Mathematical Question AnsweringMATHAccuracy53.9PHP (GPT-4 model)
Mathematical Question AnsweringSVAMPExecution Accuracy91.9GPT-4 (PHP)
Mathematical ReasoningMATHAccuracy53.9PHP (GPT-4 model)
Mathematical ReasoningSVAMPExecution Accuracy91.9GPT-4 (PHP)

Related Papers

VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17GEMMAS: Graph-based Evaluation Metrics for Multi Agent Systems2025-07-17Scaling Up RL: Unlocking Diverse Reasoning in LLMs via Prolonged Training2025-07-16DAC: A Dynamic Attention-aware Approach for Task-Agnostic Prompt Compression2025-07-16Temperature and Persona Shape LLM Agent Consensus With Minimal Accuracy Gains in Qualitative Coding2025-07-15Personalized Exercise Recommendation with Semantically-Grounded Knowledge Tracing2025-07-15KisMATH: Do LLMs Have Knowledge of Implicit Structures in Mathematical Reasoning?2025-07-15