TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/ERNIE 2.0: A Continual Pre-training Framework for Language...

ERNIE 2.0: A Continual Pre-training Framework for Language Understanding

Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Hao Tian, Hua Wu, Haifeng Wang

2019-07-29Chinese Sentiment AnalysisQuestion AnsweringChinese Reading ComprehensionSentiment AnalysisNatural Language InferenceChinese Sentence Pair ClassificationChinese Named Entity RecognitionSemantic Textual SimilarityLinguistic AcceptabilityMulti-Task LearningOpen-Domain Question AnsweringNamed Entity Recognition (NER)
PaperPDFCode(official)CodeCode

Abstract

Recently, pre-trained models have achieved state-of-the-art results in various language understanding tasks, which indicates that pre-training on large-scale corpora may play a crucial role in natural language processing. Current pre-training procedures usually focus on training the model with several simple tasks to grasp the co-occurrence of words or sentences. However, besides co-occurring, there exists other valuable lexical, syntactic and semantic information in training corpora, such as named entity, semantic closeness and discourse relations. In order to extract to the fullest extent, the lexical, syntactic and semantic information from training corpora, we propose a continual pre-training framework named ERNIE 2.0 which builds and learns incrementally pre-training tasks through constant multi-task learning. Experimental results demonstrate that ERNIE 2.0 outperforms BERT and XLNet on 16 tasks including English tasks on GLUE benchmarks and several common tasks in Chinese. The source codes and pre-trained models have been released at https://github.com/PaddlePaddle/ERNIE.

Results

TaskDatasetMetricValueModel
Question AnsweringDuReaderEM64.2ERNIE 2.0 Large
Question AnsweringDuReaderEM61.3ERNIE 2.0 Base
Natural Language InferenceWNLIAccuracy67.8ERNIE 2.0 Large
Natural Language InferenceXNLI Chinese DevAccuracy82.6ERNIE 2.0 Large
Natural Language InferenceXNLI Chinese DevAccuracy81.2ERNIE 2.0 Base
Natural Language InferenceXNLI ChineseAccuracy81ERNIE 2.0 Large
Natural Language InferenceXNLI ChineseAccuracy79.7ERNIE 2.0 Base
Natural Language InferenceMultiNLIMatched88.7ERNIE 2.0 Large
Natural Language InferenceMultiNLIMismatched88.8ERNIE 2.0 Large
Natural Language InferenceMultiNLIMatched86.1ERNIE 2.0 Base
Natural Language InferenceMultiNLIMismatched85.5ERNIE 2.0 Base
Semantic Textual SimilaritySTS BenchmarkPearson Correlation0.912ERNIE 2.0 Large
Semantic Textual SimilaritySTS BenchmarkPearson Correlation0.876ERNIE 2.0 Base
Sentiment AnalysisSST-2 Binary classificationAccuracy95ERNIE 2.0 Base
Named Entity Recognition (NER)MSRA DevF196.3ERNIE 2.0 Large
Named Entity Recognition (NER)MSRA DevF195.2ERNIE 2.0 Base
Named Entity Recognition (NER)MSRAF195ERNIE 2.0 Large
Named Entity Recognition (NER)MSRAF193.8ERNIE 2.0 Base
Open-Domain Question AnsweringDuReaderEM64.2ERNIE 2.0 Large
Open-Domain Question AnsweringDuReaderEM61.3ERNIE 2.0 Base

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17AdaptiSent: Context-Aware Adaptive Attention for Multimodal Aspect-Based Sentiment Analysis2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16