TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Contrastive Learning and Self-Training for Unsupervised Do...

Contrastive Learning and Self-Training for Unsupervised Domain Adaptation in Semantic Segmentation

Robert A. Marsden, Alexander Bartler, Mario Döbler, Bin Yang

2021-05-05Transfer LearningSemantic SegmentationSynthetic-to-Real TranslationContrastive LearningUnsupervised Domain AdaptationDomain Adaptation
PaperPDF

Abstract

Deep convolutional neural networks have considerably improved state-of-the-art results for semantic segmentation. Nevertheless, even modern architectures lack the ability to generalize well to a test dataset that originates from a different domain. To avoid the costly annotation of training data for unseen domains, unsupervised domain adaptation (UDA) attempts to provide efficient knowledge transfer from a labeled source domain to an unlabeled target domain. Previous work has mainly focused on minimizing the discrepancy between the two domains by using adversarial training or self-training. While adversarial training may fail to align the correct semantic categories as it minimizes the discrepancy between the global distributions, self-training raises the question of how to provide reliable pseudo-labels. To align the correct semantic categories across domains, we propose a contrastive learning approach that adapts category-wise centroids across domains. Furthermore, we extend our method with self-training, where we use a memory-efficient temporal ensemble to generate consistent and reliable pseudo-labels. Although both contrastive learning and self-training (CLST) through temporal ensembling enable knowledge transfer between two domains, it is their combination that leads to a symbiotic structure. We validate our approach on two domain adaptation benchmarks: GTA5 $\rightarrow$ Cityscapes and SYNTHIA $\rightarrow$ Cityscapes. Our method achieves better or comparable results than the state-of-the-art. We will make the code publicly available.

Results

TaskDatasetMetricValueModel
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU51.6CLST
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (13 classes)57.8CLST(ResNet-101)
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (16 classes)49.8CLST(ResNet-101)
Image GenerationGTAV-to-Cityscapes LabelsmIoU51.6CLST
Image GenerationSYNTHIA-to-CityscapesMIoU (13 classes)57.8CLST(ResNet-101)
Image GenerationSYNTHIA-to-CityscapesMIoU (16 classes)49.8CLST(ResNet-101)
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU51.6CLST
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (13 classes)57.8CLST(ResNet-101)
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (16 classes)49.8CLST(ResNet-101)

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17Unified Medical Image Segmentation with State Space Modeling Snake2025-07-17A Privacy-Preserving Semantic-Segmentation Method Using Domain-Adaptation Technique2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17