TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/C3: Cross-instance guided Contrastive Clustering

C3: Cross-instance guided Contrastive Clustering

Mohammadreza Sadeghi, Hadi Hojjati, Narges Armanfard

2022-11-14Deep ClusteringImage ClusteringClusteringContrastive Learning
PaperPDFCode(official)

Abstract

Clustering is the task of gathering similar data samples into clusters without using any predefined labels. It has been widely studied in machine learning literature, and recent advancements in deep learning have revived interest in this field. Contrastive clustering (CC) models are a staple of deep clustering in which positive and negative pairs of each data instance are generated through data augmentation. CC models aim to learn a feature space where instance-level and cluster-level representations of positive pairs are grouped together. Despite improving the SOTA, these algorithms ignore the cross-instance patterns, which carry essential information for improving clustering performance. This increases the false-negative-pair rate of the model while decreasing its true-positive-pair rate. In this paper, we propose a novel contrastive clustering method, Cross-instance guided Contrastive Clustering (C3), that considers the cross-sample relationships to increase the number of positive pairs and mitigate the impact of false negative, noise, and anomaly sample on the learned representation of data. In particular, we define a new loss function that identifies similar instances using the instance-level representation and encourages them to aggregate together. Moreover, we propose a novel weighting method to select negative samples in a more efficient way. Extensive experimental evaluations show that our proposed method can outperform state-of-the-art algorithms on benchmark computer vision datasets: we improve the clustering accuracy by 6.6%, 3.3%, 5.0%, 1.3% and 0.3% on CIFAR-10, CIFAR-100, ImageNet-10, ImageNet-Dogs, and Tiny-ImageNet.

Results

TaskDatasetMetricValueModel
Image ClusteringImageNet-10ARI0.861C3
Image ClusteringImageNet-10Accuracy0.942C3
Image ClusteringImageNet-10NMI0.905C3
Image ClusteringCIFAR-10ARI0.707C3
Image ClusteringCIFAR-10Accuracy0.838C3
Image ClusteringCIFAR-10NMI0.748C3
Image ClusteringTiny-ImageNetARI0.065C3
Image ClusteringTiny-ImageNetAccuracy0.141C3
Image ClusteringTiny-ImageNetNMI0.335C3
Image ClusteringCIFAR-100ARI0.275C3
Image ClusteringCIFAR-100Accuracy0.451C3
Image ClusteringCIFAR-100NMI0.434C3
Image ClusteringImagenet-dog-15ARI0.28C3
Image ClusteringImagenet-dog-15Accuracy0.434C3
Image ClusteringImagenet-dog-15NMI0.448C3

Related Papers

Tri-Learn Graph Fusion Network for Attributed Graph Clustering2025-07-18SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Ranking Vectors Clustering: Theory and Applications2025-07-16Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16LLM-Driven Dual-Level Multi-Interest Modeling for Recommendation2025-07-15