TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/An Unsupervised Sentence Embedding Method by Mutual Inform...

An Unsupervised Sentence Embedding Method by Mutual Information Maximization

Yan Zhang, Ruidan He, Zuozhu Liu, Kwan Hui Lim, Lidong Bing

2020-09-25EMNLP 2020 11Self-Supervised LearningSentence EmbeddingSentence EmbeddingsSemantic Textual SimilarityClusteringSTSSentence-Embedding
PaperPDFCode(official)

Abstract

BERT is inefficient for sentence-pair tasks such as clustering or semantic search as it needs to evaluate combinatorially many sentence pairs which is very time-consuming. Sentence BERT (SBERT) attempted to solve this challenge by learning semantically meaningful representations of single sentences, such that similarity comparison can be easily accessed. However, SBERT is trained on corpus with high-quality labeled sentence pairs, which limits its application to tasks where labeled data is extremely scarce. In this paper, we propose a lightweight extension on top of BERT and a novel self-supervised learning objective based on mutual information maximization strategies to derive meaningful sentence embeddings in an unsupervised manner. Unlike SBERT, our method is not restricted by the availability of labeled data, such that it can be applied on different domain-specific corpus. Experimental results show that the proposed method significantly outperforms other unsupervised sentence embedding baselines on common semantic textual similarity (STS) tasks and downstream supervised tasks. It also outperforms SBERT in a setting where in-domain labeled data is not available, and achieves performance competitive with supervised methods on various tasks.

Results

TaskDatasetMetricValueModel
Semantic Textual SimilaritySTS14Spearman Correlation0.6121IS-BERT-NLI
Semantic Textual SimilaritySTS15Spearman Correlation0.7523IS-BERT-NLI
Semantic Textual SimilaritySICKSpearman Correlation0.6425IS-BERT-NLI
Semantic Textual SimilaritySTS13Spearman Correlation0.6924IS-BERT-NLI
Semantic Textual SimilaritySTS BenchmarkSpearman Correlation0.6921IS-BERT-NLI
Semantic Textual SimilaritySTS12Spearman Correlation0.5677IS-BERT-NLI
Semantic Textual SimilaritySTS16Spearman Correlation0.7016IS-BERT-NLI

Related Papers

From Neurons to Semantics: Evaluating Cross-Linguistic Alignment Capabilities of Large Language Models via Neurons Alignment2025-07-20Tri-Learn Graph Fusion Network for Attributed Graph Clustering2025-07-18A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17Ranking Vectors Clustering: Theory and Applications2025-07-16Self-supervised Learning on Camera Trap Footage Yields a Strong Universal Face Embedder2025-07-14Car Object Counting and Position Estimation via Extension of the CLIP-EBC Framework2025-07-11GNN-ViTCap: GNN-Enhanced Multiple Instance Learning with Vision Transformers for Whole Slide Image Classification and Captioning2025-07-09