TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Fully Self-Supervised Learning for Semantic Segmentation

Fully Self-Supervised Learning for Semantic Segmentation

YuAn Wang, Wei Zhuo, Yucong Li, Zhi Wang, Qi Ju, Wenwu Zhu

2022-02-24Self-Supervised LearningUnsupervised Semantic SegmentationSegmentationSemantic SegmentationClustering
PaperPDF

Abstract

In this work, we present a fully self-supervised framework for semantic segmentation(FS^4). A fully bootstrapped strategy for semantic segmentation, which saves efforts for the huge amount of annotation, is crucial for building customized models from end-to-end for open-world domains. This application is eagerly needed in realistic scenarios. Even though recent self-supervised semantic segmentation methods have gained great progress, these works however heavily depend on the fully-supervised pretrained model and make it impossible a fully self-supervised pipeline. To solve this problem, we proposed a bootstrapped training scheme for semantic segmentation, which fully leveraged the global semantic knowledge for self-supervision with our proposed PGG strategy and CAE module. In particular, we perform pixel clustering and assignments for segmentation supervision. Preventing it from clustering a mess, we proposed 1) a pyramid-global-guided (PGG) training strategy to supervise the learning with pyramid image/patch-level pseudo labels, which are generated by grouping the unsupervised features. The stable global and pyramid semantic pseudo labels can prevent the segmentation from learning too many clutter regions or degrading to one background region; 2) in addition, we proposed context-aware embedding (CAE) module to generate global feature embedding in view of its neighbors close both in space and appearance in a non-trivial way. We evaluate our method on the large-scale COCO-Stuff dataset and achieved 7.19 mIoU improvements on both things and stuff objects

Results

TaskDatasetMetricValueModel
Semantic SegmentationCOCO-Stuff-27Clustering [Accuracy]40.38FS4
Semantic SegmentationCOCO-Stuff-27Clustering [mIoU]15.69FS4
Unsupervised Semantic SegmentationCOCO-Stuff-27Clustering [Accuracy]40.38FS4
Unsupervised Semantic SegmentationCOCO-Stuff-27Clustering [mIoU]15.69FS4
10-shot image generationCOCO-Stuff-27Clustering [Accuracy]40.38FS4
10-shot image generationCOCO-Stuff-27Clustering [mIoU]15.69FS4

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21Tri-Learn Graph Fusion Network for Attributed Graph Clustering2025-07-18A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17Deep Learning-Based Fetal Lung Segmentation from Diffusion-weighted MRI Images and Lung Maturity Evaluation for Fetal Growth Restriction2025-07-17DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17From Variability To Accuracy: Conditional Bernoulli Diffusion Models with Consensus-Driven Correction for Thin Structure Segmentation2025-07-17Unleashing Vision Foundation Models for Coronary Artery Segmentation: Parallel ViT-CNN Encoding and Variational Fusion2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17