TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/PiPa: Pixel- and Patch-wise Self-supervised Learning for D...

PiPa: Pixel- and Patch-wise Self-supervised Learning for Domain Adaptative Semantic Segmentation

Mu Chen, Zhedong Zheng, Yi Yang, Tat-Seng Chua

2022-11-14Self-Supervised LearningSemantic SegmentationSynthetic-to-Real TranslationUnsupervised Domain AdaptationImage-to-Image TranslationDomain Adaptation
PaperPDFCode(official)

Abstract

Unsupervised Domain Adaptation (UDA) aims to enhance the generalization of the learned model to other domains. The domain-invariant knowledge is transferred from the model trained on labeled source domain, e.g., video game, to unlabeled target domains, e.g., real-world scenarios, saving annotation expenses. Existing UDA methods for semantic segmentation usually focus on minimizing the inter-domain discrepancy of various levels, e.g., pixels, features, and predictions, for extracting domain-invariant knowledge. However, the primary intra-domain knowledge, such as context correlation inside an image, remains underexplored. In an attempt to fill this gap, we propose a unified pixel- and patch-wise self-supervised learning framework, called PiPa, for domain adaptive semantic segmentation that facilitates intra-image pixel-wise correlations and patch-wise semantic consistency against different contexts. The proposed framework exploits the inherent structures of intra-domain images, which: (1) explicitly encourages learning the discriminative pixel-wise features with intra-class compactness and inter-class separability, and (2) motivates the robust feature learning of the identical patch against different contexts or fluctuations. Extensive experiments verify the effectiveness of the proposed method, which obtains competitive accuracy on the two widely-used UDA benchmarks, i.e., 75.6 mIoU on GTA to Cityscapes and 68.2 mIoU on Synthia to Cityscapes. Moreover, our method is compatible with other UDA approaches to further improve the performance without introducing extra parameters.

Results

TaskDatasetMetricValueModel
Image-to-Image TranslationSYNTHIA-to-CityscapesmIoU (13 classes)74.8HRDA + PiPa
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU71.7DAFormer + PiPa
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU75.6HRDA+PiPa
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU71.7DAFormer+PiPa
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (13 classes)74.8HRDA+PiPa
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (16 classes)68.2HRDA+PiPa
Domain AdaptationSYNTHIA-to-CityscapesmIoU68.2HRDA+PiPa
Domain AdaptationGTA5 to CityscapesmIoU75.6HRDA+PiPa
Domain AdaptationGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
Domain AdaptationGTAV-to-Cityscapes LabelsmIoU71.7DAFormer + PiPa
Domain AdaptationSYNTHIA-to-CityscapesmIoU (13 classes)74.8HRDA + PiPa
Image GenerationSYNTHIA-to-CityscapesmIoU (13 classes)74.8HRDA + PiPa
Image GenerationGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
Image GenerationGTAV-to-Cityscapes LabelsmIoU71.7DAFormer + PiPa
Image GenerationGTAV-to-Cityscapes LabelsmIoU75.6HRDA+PiPa
Image GenerationGTAV-to-Cityscapes LabelsmIoU71.7DAFormer+PiPa
Image GenerationSYNTHIA-to-CityscapesMIoU (13 classes)74.8HRDA+PiPa
Image GenerationSYNTHIA-to-CityscapesMIoU (16 classes)68.2HRDA+PiPa
Semantic SegmentationGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
Semantic SegmentationSYNTHIA-to-CityscapesMean IoU68.2HRDA + PiPa
Unsupervised Domain AdaptationGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
Unsupervised Domain AdaptationGTAV-to-Cityscapes LabelsmIoU71.7DAFormer + PiPa
Unsupervised Domain AdaptationSYNTHIA-to-CityscapesmIoU (13 classes)74.8HRDA + PiPa
10-shot image generationGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
10-shot image generationSYNTHIA-to-CityscapesMean IoU68.2HRDA + PiPa
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesmIoU (13 classes)74.8HRDA + PiPa
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU75.6HRDA + PiPa
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU71.7DAFormer + PiPa
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU75.6HRDA+PiPa
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU71.7DAFormer+PiPa
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (13 classes)74.8HRDA+PiPa
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (16 classes)68.2HRDA+PiPa

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17Unified Medical Image Segmentation with State Space Modeling Snake2025-07-17A Privacy-Preserving Semantic-Segmentation Method Using Domain-Adaptation Technique2025-07-17SAMST: A Transformer framework based on SAM pseudo label filtering for remote sensing semi-supervised semantic segmentation2025-07-16Tomato Multi-Angle Multi-Pose Dataset for Fine-Grained Phenotyping2025-07-15