TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/LinkBERT: Pretraining Language Models with Document Links

LinkBERT: Pretraining Language Models with Document Links

Michihiro Yasunaga, Jure Leskovec, Percy Liang

2022-03-29ACL 2022 5Text ClassificationQuestion AnsweringRelation ExtractionSentence SimilarityMasked Language ModelingTriviaQASemantic SimilarityDocument ClassificationRelation PredictionMedical Relation ExtractionNamed Entity Recognition (NER)Language Modelling
PaperPDFCode(official)

Abstract

Language model (LM) pretraining can learn various knowledge from text corpora, helping downstream tasks. However, existing methods such as BERT model a single document, and do not capture dependencies or knowledge that span across documents. In this work, we propose LinkBERT, an LM pretraining method that leverages links between documents, e.g., hyperlinks. Given a text corpus, we view it as a graph of documents and create LM inputs by placing linked documents in the same context. We then pretrain the LM with two joint self-supervised objectives: masked language modeling and our new proposal, document relation prediction. We show that LinkBERT outperforms BERT on various downstream tasks across two domains: the general domain (pretrained on Wikipedia with hyperlinks) and biomedical domain (pretrained on PubMed with citation links). LinkBERT is especially effective for multi-hop reasoning and few-shot QA (+5% absolute improvement on HotpotQA and TriviaQA), and our biomedical LinkBERT sets new states of the art on various BioNLP tasks (+7% on BioASQ and USMLE). We release our pretrained models, LinkBERT and BioLinkBERT, as well as code and data at https://github.com/michiyasunaga/LinkBERT.

Results

TaskDatasetMetricValueModel
Relation ExtractionGADF184.9BioLinkBERT (large)
Relation ExtractionGADMicro F184.9BioLinkBERT (large)
Relation ExtractionDDIF183.35BioLinkBERT (large)
Relation ExtractionDDIMicro F183.35BioLinkBERT (large)
Relation ExtractionChemProtF179.98BioLinkBERT (large)
Relation ExtractionChemProtMicro F179.98BioLinkBERT (large)
Question AnsweringMRQAAverage F181LinkBERT (large)
Question AnsweringBLURBAccuracy83.5BioLinkBERT (large)
Question AnsweringBLURBAccuracy80.81BioLinkBERT (base)
Question AnsweringPubMedQAAccuracy72.2BioLinkBERT (large)
Question AnsweringPubMedQAAccuracy70.2BioLinkBERT (base)
Question AnsweringMedQAAccuracy40BioLinkBERT (base)
Question AnsweringBioASQAccuracy94.8BioLinkBERT (large)
Question AnsweringBioASQAccuracy91.4BioLinkBERT (base)
Question AnsweringNewsQAF172.6LinkBERT (large)
Question AnsweringSQuAD1.1EM87.45LinkBERT (large)
Question AnsweringSQuAD1.1F192.7LinkBERT (large)
Question AnsweringTriviaQAF178.2LinkBERT (large)
Language ModellingBIOSSESPearson Correlation0.9363BioLinkBERT (large)
Language ModellingBIOSSESPearson Correlation0.9325BioLinkBERT (base)
Medical Relation ExtractionDDI extraction 2013 corpusF183.35BioLinkBERT (large)
Named Entity Recognition (NER)NCBI-diseaseF188.76BioLinkBERT (large)
Named Entity Recognition (NER)BC5CDR-chemicalF194.04BioLinkBERT (large)
Named Entity Recognition (NER)BC5CDR-diseaseF186.39BioLinkBERT (large)
Named Entity Recognition (NER)BC2GMF185.18BioLinkBERT (large)
Named Entity Recognition (NER)BC5CDRF190.22BioLinkBERT (large)
Named Entity Recognition (NER)JNLPBAF180.06BioLinkBERT (large)
Text ClassificationBLURBF184.88BioLinkBERT (large)
Text ClassificationBLURBF184.35BioLinkBERT (base)
Text ClassificationHOCF188.1BioLinkBERT (large)
Text ClassificationHOCMicro F184.87BioLinkBERT (large)
Sentence Pair ModelingBIOSSESPearson Correlation0.9363BioLinkBERT (large)
Sentence Pair ModelingBIOSSESPearson Correlation0.9325BioLinkBERT (base)
Document ClassificationHOCF188.1BioLinkBERT (large)
Document ClassificationHOCMicro F184.87BioLinkBERT (large)
Biomedical Information RetrievalEBM PICOMacro F1 word level74.19BioLinkBERT (large)
Biomedical Information RetrievalEBM PICOMacro F1 word level73.97BioLinkBERT (base)
ClassificationBLURBF184.88BioLinkBERT (large)
ClassificationBLURBF184.35BioLinkBERT (base)
ClassificationHOCF188.1BioLinkBERT (large)
ClassificationHOCMicro F184.87BioLinkBERT (large)
Semantic SimilarityBIOSSESPearson Correlation0.9363BioLinkBERT (large)
Semantic SimilarityBIOSSESPearson Correlation0.9325BioLinkBERT (base)

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Making Language Model a Hierarchical Classifier and Generator2025-07-17From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17