TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Sentence-BERT: Sentence Embeddings using Siamese BERT-Netw...

Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks

Nils Reimers, Iryna Gurevych

2019-08-27IJCNLP 2019 11Sentence EmbeddingTransfer LearningLinear-Probe ClassificationSentence EmbeddingsSemantic SimilaritySemantic Textual SimilarityClusteringSTS
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

BERT (Devlin et al., 2018) and RoBERTa (Liu et al., 2019) has set a new state-of-the-art performance on sentence-pair regression tasks like semantic textual similarity (STS). However, it requires that both sentences are fed into the network, which causes a massive computational overhead: Finding the most similar pair in a collection of 10,000 sentences requires about 50 million inference computations (~65 hours) with BERT. The construction of BERT makes it unsuitable for semantic similarity search as well as for unsupervised tasks like clustering. In this publication, we present Sentence-BERT (SBERT), a modification of the pretrained BERT network that use siamese and triplet network structures to derive semantically meaningful sentence embeddings that can be compared using cosine-similarity. This reduces the effort for finding the most similar pair from 65 hours with BERT / RoBERTa to about 5 seconds with SBERT, while maintaining the accuracy from BERT. We evaluate SBERT and SRoBERTa on common STS tasks and transfer learning tasks, where it outperforms other state-of-the-art sentence embeddings methods.

Results

TaskDatasetMetricValueModel
Semantic Textual SimilaritySTS14Spearman Correlation0.7490000000000001SBERT-NLI-large
Semantic Textual SimilaritySTS15Spearman Correlation0.8185SRoBERTa-NLI-large
Semantic Textual SimilaritySICKSpearman Correlation0.7462SentenceBERT
Semantic Textual SimilaritySICKSpearman Correlation0.7446SRoBERTa-NLI-base
Semantic Textual SimilaritySICKSpearman Correlation0.7429SRoBERTa-NLI-large
Semantic Textual SimilaritySICKSpearman Correlation0.7375SBERT-NLI-large
Semantic Textual SimilaritySICKSpearman Correlation0.7291SBERT-NLI-base
Semantic Textual SimilaritySTS13Spearman Correlation0.7846SBERT-NLI-large
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.8615SRoBERTa-NLI-STSb-large
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.8479SBERT-STSb-base
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.8445SBERT-STSb-large
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.79SBERT-NLI-large
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.7777SRoBERTa-NLI-base
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.7703SBERT-NLI-base
Semantic Textual SimilaritySTS12Spearman Correlation0.7453SRoBERTa-NLI-large
Semantic Textual SimilaritySTS16Spearman Correlation0.7682SRoBERTa-NLI-large

Related Papers

From Neurons to Semantics: Evaluating Cross-Linguistic Alignment Capabilities of Large Language Models via Neurons Alignment2025-07-20RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Tri-Learn Graph Fusion Network for Attributed Graph Clustering2025-07-18Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows2025-07-16Ranking Vectors Clustering: Theory and Applications2025-07-16Robust-Multi-Task Gradient Boosting2025-07-15