Non-contrastive sentence representations via self-supervision
Marco Farina, Duccio Pappadopulo
2023-10-26Sentence Embeddings
Abstract
Sample contrastive methods, typically referred to simply as contrastive are the foundation of most unsupervised methods to learn text and sentence embeddings. On the other hand, a different class of self-supervised loss functions and methods have been considered in the computer vision community and referred to as dimension contrastive. In this paper, we thoroughly compare this class of methods with the standard baseline for contrastive sentence embeddings, SimCSE. We find that self-supervised embeddings trained using dimension contrastive objectives can outperform SimCSE on downstream tasks without needing auxiliary loss functions.
Related Papers
From Neurons to Semantics: Evaluating Cross-Linguistic Alignment Capabilities of Large Language Models via Neurons Alignment2025-07-20SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17Intrinsic vs. Extrinsic Evaluation of Czech Sentence Embeddings: Semantic Relevance Doesn't Help with MT Evaluation2025-06-25Do We Talk to Robots Like Therapists, and Do They Respond Accordingly? Language Alignment in AI Emotional Support2025-06-19Mechanistic Decomposition of Sentence Representations2025-06-04Rethinking the Understanding Ability across LLMs through Mutual Information2025-05-25LLMs Are Not Scorers: Rethinking MT Evaluation with Generation-Based Methods2025-05-22Contrastive Prompting Enhances Sentence Embeddings in LLMs through Inference-Time Steering2025-05-19