TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Control Prefixes for Parameter-Efficient Text Generation

Control Prefixes for Parameter-Efficient Text Generation

Jordan Clive, Kris Cao, Marek Rei

2021-10-15Data-to-Text GenerationText GenerationAttributeAbstractive Text SummarizationText SummarizationText SimplificationLanguage Modelling
PaperPDFCode(official)Code(official)

Abstract

Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to a downstream application. However, it uses the same dataset-level tuned prompt for all examples in the dataset. We extend this idea and propose a dynamic method, Control Prefixes, which allows for the inclusion of conditional input-dependent information, combining the benefits of prompt tuning and controlled generation. The method incorporates attribute-level learnable representations into different layers of a pre-trained transformer, allowing for the generated text to be guided in a particular direction. We provide a systematic evaluation of the technique and apply it to five datasets from the GEM benchmark for natural language generation (NLG). Although the aim is to develop a parameter-efficient model, we show Control Prefixes can even outperform full fine-tuning methods. We present state-of-the-art results on several data-to-text datasets, including WebNLG.

Results

TaskDatasetMetricValueModel
Text GenerationDARTMETEOR0.411Control Prefixes (T5-large)
Text GenerationWebNLGBLEU67.32Control Prefixes (A1, T5-large)
Text GenerationWebNLGBLEU67.15Control Prefixes (A1, A2, T5-large)
Text GenerationCleaned E2E NLG ChallengeBLEU (Test set)44.15Control Prefixes (T5-large)
Text GenerationWebNLG FullBLEU62.27Control Prefixes (A1, A2, T5-large)
Text GenerationWebNLG FullBLEU61.94Control Prefixes (A1, T5-large)
Text SimplificationTurkCorpusFKGL7.74Control Prefixes (BART)
Text SimplificationTurkCorpusQuestEval (Reference-less, BERTScore)0.66Control Prefixes (BART)
Text SimplificationTurkCorpusSARI (EASSE>=0.2.1)42.32Control Prefixes (BART)
Text SimplificationASSETFKGL5.97Control Prefixes (BART)
Text SimplificationASSETQuestEval (Reference-less, BERTScore)0.64Control Prefixes (BART)
Text SimplificationASSETSARI (EASSE>=0.2.1)43.58Control Prefixes (BART)
Data-to-Text GenerationWebNLGBLEU67.32Control Prefixes (A1, T5-large)
Data-to-Text GenerationWebNLGBLEU67.15Control Prefixes (A1, A2, T5-large)
Data-to-Text GenerationCleaned E2E NLG ChallengeBLEU (Test set)44.15Control Prefixes (T5-large)
Data-to-Text GenerationWebNLG FullBLEU62.27Control Prefixes (A1, A2, T5-large)
Data-to-Text GenerationWebNLG FullBLEU61.94Control Prefixes (A1, T5-large)

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16MGFFD-VLM: Multi-Granularity Prompt Learning for Face Forgery Detection with VLM2025-07-16Non-Adaptive Adversarial Face Generation2025-07-16