TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CEDR: Contextualized Embeddings for Document Ranking

CEDR: Contextualized Embeddings for Document Ranking

Sean MacAvaney, Andrew Yates, Arman Cohan, Nazli Goharian

2019-04-15Ad-Hoc Information RetrievalGeneral ClassificationDocument Ranking
PaperPDFCodeCode(official)CodeCode(official)CodeCodeCode

Abstract

Although considerable attention has been given to neural ranking architectures recently, far less attention has been paid to the term representations that are used as input to these models. In this work, we investigate how two pretrained contextualized language models (ELMo and BERT) can be utilized for ad-hoc document ranking. Through experiments on TREC benchmarks, we find that several existing neural ranking architectures can benefit from the additional context provided by contextualized language models. Furthermore, we propose a joint approach that incorporates BERT's classification vector into existing neural models and show that it outperforms state-of-the-art ad-hoc ranking baselines. We call this joint approach CEDR (Contextualized Embeddings for Document Ranking). We also address practical challenges in using these models for ranking, including the maximum input length imposed by BERT and runtime performance impacts of contextualized language models.

Results

TaskDatasetMetricValueModel
Ad-Hoc Information RetrievalTREC Robust04P@200.4667CEDR-KNRM
Ad-Hoc Information RetrievalTREC Robust04nDCG@200.5381CEDR-KNRM
Ad-Hoc Information RetrievalTREC Robust04P@200.4042Vanilla BERT
Ad-Hoc Information RetrievalTREC Robust04nDCG@200.4541Vanilla BERT

Related Papers

Precise Zero-Shot Pointwise Ranking with LLMs through Post-Aggregated Global Context Information2025-06-12Bridge the Gap between Past and Future: Siamese Model Optimization for Context-Aware Document Ranking2025-05-20Beyond Retrieval: Joint Supervision and Multimodal Document Ranking for Textbook Question Answering2025-05-17A Unified Retrieval Framework with Document Ranking and EDU Filtering for Multi-document Summarization2025-04-23Specialized text classification: an approach to classifying Open Banking transactions2025-04-10How do Large Language Models Understand Relevance? A Mechanistic Interpretability Perspective2025-04-10Beyond Reproducibility: Advancing Zero-shot LLM Reranking Efficiency with Setwise Insertion2025-04-09Distillation and Refinement of Reasoning in Small Language Models for Document Re-ranking2025-04-04