TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Neural sentence embedding models for semantic similarity e...

Neural sentence embedding models for semantic similarity estimation in the biomedical domain

Kathrin Blagec, Hong Xu, Asan Agibetov, Matthias Samwald

2021-10-01Sentence Embeddings For Biomedical TextsSentence EmbeddingSemantic SimilaritySemantic Textual SimilaritySentence-Embedding
PaperPDFCode(official)

Abstract

BACKGROUND: In this study, we investigated the efficacy of current state-of-the-art neural sentence embedding models for semantic similarity estimation of sentences from biomedical literature. We trained different neural embedding models on 1.7 million articles from the PubMed Open Access dataset, and evaluated them based on a biomedical benchmark set containing 100 sentence pairs annotated by human experts and a smaller contradiction subset derived from the original benchmark set. RESULTS: With a Pearson correlation of 0.819, our best unsupervised model based on the Paragraph Vector Distributed Memory algorithm outperforms previous state-of-the-art results achieved on the BIOSSES biomedical benchmark set. Moreover, our proposed supervised model that combines different string-based similarity metrics with a neural embedding model surpasses previous ontology-dependent supervised state-of-the-art approaches in terms of Pearson's r (r=0.871) on the biomedical benchmark set. In contrast to the promising results for the original benchmark, we found our best models' performance on the smaller contradiction subset to be poor. CONCLUSIONS: In this study we highlighted the value of neural network-based models for semantic similarity estimation in the biomedical domain by showing that they can keep up with and even surpass previous state-of-the-art approaches for semantic similarity estimation that depend on the availability of laboriously curated ontologies when evaluated on a biomedical benchmark set. Capturing contradictions and negations in biomedical sentences, however, emerged as an essential area for further work.

Results

TaskDatasetMetricValueModel
Sentence EmbeddingsBIOSSESPearson Correlation0.871Supervised combination of: Jaccard, Q-gram, sent2vec, Paragraph vector DM, skip-thoughts, fastText
Sentence EmbeddingsBIOSSESPearson Correlation0.846Unsupervised combination (mean) of: Jaccard, q-gram, Paragraph vector (PV-DBOW) and sent2vec
Sentence EmbeddingsBIOSSESPearson Correlation0.819Paragraph vector (PV-DM)
Sentence EmbeddingsBIOSSESPearson Correlation0.804Paragraph vector (PV-DBOW)
Sentence EmbeddingsBIOSSESPearson Correlation0.798Sent2vec
Sentence EmbeddingsBIOSSESPearson Correlation0.766fastText (skip-gram, max pooling)
Sentence EmbeddingsBIOSSESPearson Correlation0.723Q-gram (q = 3)
Sentence EmbeddingsBIOSSESPearson Correlation0.485Skip-thoughts
Sentence EmbeddingsBIOSSESPearson Correlation0.253fastText (CBOW, max pooling)
Representation LearningBIOSSESPearson Correlation0.871Supervised combination of: Jaccard, Q-gram, sent2vec, Paragraph vector DM, skip-thoughts, fastText
Representation LearningBIOSSESPearson Correlation0.846Unsupervised combination (mean) of: Jaccard, q-gram, Paragraph vector (PV-DBOW) and sent2vec
Representation LearningBIOSSESPearson Correlation0.819Paragraph vector (PV-DM)
Representation LearningBIOSSESPearson Correlation0.804Paragraph vector (PV-DBOW)
Representation LearningBIOSSESPearson Correlation0.798Sent2vec
Representation LearningBIOSSESPearson Correlation0.766fastText (skip-gram, max pooling)
Representation LearningBIOSSESPearson Correlation0.723Q-gram (q = 3)
Representation LearningBIOSSESPearson Correlation0.485Skip-thoughts
Representation LearningBIOSSESPearson Correlation0.253fastText (CBOW, max pooling)

Related Papers

SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17SARA: Selective and Adaptive Retrieval-augmented Generation with Context Compression2025-07-08FA: Forced Prompt Learning of Vision-Language Models for Out-of-Distribution Detection2025-07-06LineRetriever: Planning-Aware Observation Reduction for Web Agents2025-06-30Enhancing Automatic Term Extraction with Large Language Models via Syntactic Retrieval2025-06-26DALR: Dual-level Alignment Learning for Multimodal Sentence Representation Learning2025-06-26Intrinsic vs. Extrinsic Evaluation of Czech Sentence Embeddings: Semantic Relevance Doesn't Help with MT Evaluation2025-06-25Leveraging Vision-Language Models to Select Trustworthy Super-Resolution Samples Generated by Diffusion Models2025-06-25