TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Unified Language Model Pre-training for Natural Language U...

Unified Language Model Pre-training for Natural Language Understanding and Generation

Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon

2019-05-08NeurIPS 2019 12Question AnsweringText GenerationAbstractive Text SummarizationText SummarizationGenerative Question AnsweringNatural Language UnderstandingDocument SummarizationQuestion GenerationLanguage ModellingResponse Generation
PaperPDFCodeCodeCodeCodeCodeCodeCode(official)CodeCode

Abstract

This paper presents a new Unified pre-trained Language Model (UniLM) that can be fine-tuned for both natural language understanding and generation tasks. The model is pre-trained using three types of language modeling tasks: unidirectional, bidirectional, and sequence-to-sequence prediction. The unified modeling is achieved by employing a shared Transformer network and utilizing specific self-attention masks to control what context the prediction conditions on. UniLM compares favorably with BERT on the GLUE benchmark, and the SQuAD 2.0 and CoQA question answering tasks. Moreover, UniLM achieves new state-of-the-art results on five natural language generation datasets, including improving the CNN/DailyMail abstractive summarization ROUGE-L to 40.51 (2.04 absolute improvement), the Gigaword abstractive summarization ROUGE-L to 35.75 (0.86 absolute improvement), the CoQA generative question answering F1 score to 82.5 (37.1 absolute improvement), the SQuAD question generation BLEU-4 to 22.12 (3.75 absolute improvement), and the DSTC7 document-grounded dialog response generation NIST-4 to 2.67 (human performance is 2.65). The code and pre-trained models are available at https://github.com/microsoft/unilm.

Results

TaskDatasetMetricValueModel
Question AnsweringCoQAF1-Score82.5UniLM
Text SummarizationGigaWordROUGE-138.9UniLM
Text SummarizationGigaWordROUGE-220.05UniLM
Text SummarizationGigaWordROUGE-L36UniLM
Text SummarizationCNN / Daily MailROUGE-143.08UniLM
Text SummarizationCNN / Daily MailROUGE-220.43UniLM
Text SummarizationCNN / Daily MailROUGE-L40.34UniLM
Text SummarizationCNN / Daily MailROUGE-143.08UniLM (Abstractive Summarization)
Text SummarizationCNN / Daily MailROUGE-220.43UniLM (Abstractive Summarization)
Text SummarizationCNN / Daily MailROUGE-L40.34UniLM (Abstractive Summarization)
Abstractive Text SummarizationCNN / Daily MailROUGE-143.08UniLM
Abstractive Text SummarizationCNN / Daily MailROUGE-220.43UniLM
Abstractive Text SummarizationCNN / Daily MailROUGE-L40.34UniLM
Question GenerationSQuAD1.1BLEU-422.78UniLM
Question GenerationSQuAD1.1METEOR25.1UniLM
Question GenerationSQuAD1.1ROUGE-L51.1UniLM
Document SummarizationCNN / Daily MailROUGE-143.08UniLM (Abstractive Summarization)
Document SummarizationCNN / Daily MailROUGE-220.43UniLM (Abstractive Summarization)
Document SummarizationCNN / Daily MailROUGE-L40.34UniLM (Abstractive Summarization)

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17