TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MV-MR: multi-views and multi-representations for self-supe...

MV-MR: multi-views and multi-representations for self-supervised learning and knowledge distillation

Vitaliy Kinakh, Mariia Drozdova, Slava Voloshynovskiy

2023-03-21Self-Supervised Image ClassificationSelf-Supervised LearningClusteringContrastive LearningKnowledge DistillationUnsupervised Image ClassificationLinear evaluation
PaperPDFCode(official)

Abstract

We present a new method of self-supervised learning and knowledge distillation based on the multi-views and multi-representations (MV-MR). The MV-MR is based on the maximization of dependence between learnable embeddings from augmented and non-augmented views, jointly with the maximization of dependence between learnable embeddings from augmented view and multiple non-learnable representations from non-augmented view. We show that the proposed method can be used for efficient self-supervised classification and model-agnostic knowledge distillation. Unlike other self-supervised techniques, our approach does not use any contrastive learning, clustering, or stop gradients. MV-MR is a generic framework allowing the incorporation of constraints on the learnable embeddings via the usage of image multi-representations as regularizers. Along this line, knowledge distillation is considered a particular case of such a regularization. MV-MR provides the state-of-the-art performance on the STL10 and ImageNet-1K datasets among non-contrastive and clustering-free methods. We show that a lower complexity ResNet50 model pretrained using proposed knowledge distillation based on the CLIP ViT model achieves state-of-the-art performance on STL10 linear evaluation. The code is available at: https://github.com/vkinakh/mv-mr

Results

TaskDatasetMetricValueModel
Image ClassificationImageNetTop 5 Accuracy92.1MV-MR
Image ClassificationSTL-10Accuracy89.67MV-MR
Image ClassificationCIFAR-20Accuracy73.2MV-MR
Self-Supervised LearningSTL-10Accuracy89.67MV-MR
Knowledge DistillationCIFAR-100Top-1 Accuracy (%)78.6MV-MR (T: CLIP/ViT-B-16 S: resnet50)

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Tri-Learn Graph Fusion Network for Attributed Graph Clustering2025-07-18A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17