TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Revisiting Weakly Supervised Pre-Training of Visual Percep...

Revisiting Weakly Supervised Pre-Training of Visual Perception Models

Mannat Singh, Laura Gustafson, Aaron Adcock, Vinicius de Freitas Reis, Bugra Gedik, Raj Prateek Kosaraju, Dhruv Mahajan, Ross Girshick, Piotr Dollár, Laurens van der Maaten

2022-01-20CVPR 2022 1Image ClassificationSelf-Supervised LearningTransfer LearningOut-of-Distribution GeneralizationFine-Grained Image Classification
PaperPDFCode(official)Code

Abstract

Model pre-training is a cornerstone of modern visual recognition systems. Although fully supervised pre-training on datasets like ImageNet is still the de-facto standard, recent studies suggest that large-scale weakly supervised pre-training can outperform fully supervised approaches. This paper revisits weakly-supervised pre-training of models using hashtag supervision with modern versions of residual networks and the largest-ever dataset of images and corresponding hashtags. We study the performance of the resulting models in various transfer-learning settings including zero-shot transfer. We also compare our models with those obtained via large-scale self-supervised learning. We find our weakly-supervised models to be very competitive across all settings, and find they substantially outperform their self-supervised counterparts. We also include an investigation into whether our models learned potentially troubling associations or stereotypes. Overall, our results provide a compelling argument for the use of weakly supervised learning in the development of visual recognition systems. Our models, Supervised Weakly through hashtAGs (SWAG), are available publicly.

Results

TaskDatasetMetricValueModel
Image ClassificationImageNet V2Top 1 Accuracy81.1SWAG (ViT H/14)
Image ClassificationPlaces365-StandardTop 1 Accuracy60.7SWAG (ViT H/14)
Image ClassificationObjectNetTop-1 Accuracy69.5SWAG (ViT H/14)
Image ClassificationObjectNetTop-1 Accuracy64.3RegNetY 128GF (Platt)
Image ClassificationObjectNetTop-1 Accuracy60ViT H/14 (Platt)
Image ClassificationObjectNetTop-1 Accuracy57.3ViT L/16 (Platt)
Image ClassificationObjectNetTop-1 Accuracy48.9ViT B/16
Image ClassificationImageNetGFLOPs1018.8SWAG (ViT H/14)
Image ClassificationCUB-200-2011Accuracy91.7SWAG (ViT H/14)
Fine-Grained Image ClassificationCUB-200-2011Accuracy91.7SWAG (ViT H/14)

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17