TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SSL4EO-S12: A Large-Scale Multi-Modal, Multi-Temporal Data...

SSL4EO-S12: A Large-Scale Multi-Modal, Multi-Temporal Dataset for Self-Supervised Learning in Earth Observation

Yi Wang, Nassim Ait Ali Braham, Zhitong Xiong, Chenying Liu, Conrad M Albrecht, Xiao Xiang Zhu

2022-11-13Multi-Label Image ClassificationSelf-Supervised Learning
PaperPDFCodeCodeCode(official)Code

Abstract

Self-supervised pre-training bears potential to generate expressive representations without human annotation. Most pre-training in Earth observation (EO) are based on ImageNet or medium-size, labeled remote sensing (RS) datasets. We share an unlabeled RS dataset SSL4EO-S12 (Self-Supervised Learning for Earth Observation - Sentinel-1/2) to assemble a large-scale, global, multimodal, and multi-seasonal corpus of satellite imagery from the ESA Sentinel-1 \& -2 satellite missions. For EO applications we demonstrate SSL4EO-S12 to succeed in self-supervised pre-training for a set of methods: MoCo-v2, DINO, MAE, and data2vec. Resulting models yield downstream performance close to, or surpassing accuracy measures of supervised learning. In addition, pre-training on SSL4EO-S12 excels compared to existing datasets. We make openly available the dataset, related source code, and pre-trained models at https://github.com/zhu-xlab/SSL4EO-S12.

Results

TaskDatasetMetricValueModel
Multi-Label Image ClassificationBigEarthNet (official test set)F1 Score80.5MoCov3 (ViT-S/16)
Multi-Label Image ClassificationBigEarthNet (official test set)mAP (micro)89.3MoCov3 (ViT-S/16)
Multi-Label Image ClassificationBigEarthNet (official test set)F1 Score79.8MoCov2 (ResNet50)
Multi-Label Image ClassificationBigEarthNet (official test set)mAP (micro)88.7MoCov2 (ResNet50)
Multi-Label Image ClassificationBigEarthNetmAP (micro)91.8MoCo-v2 (ResNet50, fine tune)
Multi-Label Image ClassificationBigEarthNetmAP (micro)89.9MoCo-v3 (ViT-S/16, fine tune)
Multi-Label Image ClassificationBigEarthNetmAP (micro)88.9MAE (ViT-S/16, fine tune)
Image ClassificationBigEarthNet (official test set)F1 Score80.5MoCov3 (ViT-S/16)
Image ClassificationBigEarthNet (official test set)mAP (micro)89.3MoCov3 (ViT-S/16)
Image ClassificationBigEarthNet (official test set)F1 Score79.8MoCov2 (ResNet50)
Image ClassificationBigEarthNet (official test set)mAP (micro)88.7MoCov2 (ResNet50)
Image ClassificationBigEarthNetmAP (micro)91.8MoCo-v2 (ResNet50, fine tune)
Image ClassificationBigEarthNetmAP (micro)89.9MoCo-v3 (ViT-S/16, fine tune)
Image ClassificationBigEarthNetmAP (micro)88.9MAE (ViT-S/16, fine tune)

Related Papers

A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17Self-supervised Learning on Camera Trap Footage Yields a Strong Universal Face Embedder2025-07-14Speech Quality Assessment Model Based on Mixture of Experts: System-Level Performance Enhancement and Utterance-Level Challenge Analysis2025-07-08World4Drive: End-to-End Autonomous Driving via Intention-aware Physical Latent World Model2025-07-01ShapeEmbed: a self-supervised learning framework for 2D contour quantification2025-07-01RetFiner: A Vision-Language Refinement Scheme for Retinal Foundation Models2025-06-27Boosting Generative Adversarial Transferability with Self-supervised Vision Transformer Features2025-06-26Hybrid Deep Learning and Signal Processing for Arabic Dialect Recognition in Low-Resource Settings2025-06-26