TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Metric-Based In-context Learning: A Case Study in Text Sim...

Metric-Based In-context Learning: A Case Study in Text Simplification

Subha Vadlamannati, Gözde Gül Şahin

2023-07-27Text Simplification
PaperPDFCode(official)

Abstract

In-context learning (ICL) for large language models has proven to be a powerful approach for many natural language processing tasks. However, determining the best method to select examples for ICL is nontrivial as the results can vary greatly depending on the quality, quantity, and order of examples used. In this paper, we conduct a case study on text simplification (TS) to investigate how to select the best and most robust examples for ICL. We propose Metric-Based in-context Learning (MBL) method that utilizes commonly used TS metrics such as SARI, compression ratio, and BERT-Precision for selection. Through an extensive set of experiments with various-sized GPT models on standard TS benchmarks such as TurkCorpus and ASSET, we show that examples selected by the top SARI scores perform the best on larger models such as GPT-175B, while the compression ratio generally performs better on smaller models such as GPT-13B and GPT-6.7B. Furthermore, we demonstrate that MBL is generally robust to example orderings and out-of-domain test sets, and outperforms strong baselines and state-of-the-art finetuned language models. Finally, we show that the behaviour of large GPT models can be implicitly controlled by the chosen metric. Our research provides a new framework for selecting examples in ICL, and demonstrates its effectiveness in text simplification tasks, breaking new ground for more accurate and efficient NLG systems.

Results

TaskDatasetMetricValueModel
Text SimplificationTurkCorpusBLEU79.83GPT-175B (6 SARI-selected examples, high/low)
Text SimplificationTurkCorpusFKGL9.33GPT-175B (6 SARI-selected examples, high/low)
Text SimplificationTurkCorpusSARI (EASSE>=0.2.1)43.46GPT-175B (6 SARI-selected examples, high/low)
Text SimplificationASSETBLEU73.92GPT-175B (15 SARI-selected examples, random ordering)
Text SimplificationASSETFKGL7.73GPT-175B (15 SARI-selected examples, random ordering)
Text SimplificationASSETSARI (EASSE>=0.2.1)47.94GPT-175B (15 SARI-selected examples, random ordering)

Related Papers

Document-Level Text Generation with Minimum Bayes Risk Decoding using Optimal Transport2025-05-29Enhancing Paraphrase Type Generation: The Impact of DPO and RLHF Evaluated with Human-Ranked Data2025-05-28Automated Feedback Loops to Protect Text Simplification with Generative AI from Information Loss2025-05-22Resource for Error Analysis in Text Simplification: New Taxonomy and Test Collection2025-05-22A Structured Literature Review on Traditional Approaches in Current Natural Language Processing2025-05-19LLM-based Text Simplification and its Effect on User Comprehension and Cognitive Load2025-05-04SimplifyMyText: An LLM-Based System for Inclusive Plain Language Text Simplification2025-04-19Evaluation Under Imperfect Benchmarks and Ratings: A Case Study in Text Simplification2025-04-13