TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Exploring High-quality Target Domain Information for Unsup...

Exploring High-quality Target Domain Information for Unsupervised Domain Adaptive Semantic Segmentation

Junjie Li, Zilei Wang, Yuan Gao, Xiaoming Hu

2022-08-12Semantic SegmentationSynthetic-to-Real TranslationContrastive LearningImage-to-Image TranslationDomain Adaptation
PaperPDFCode(official)

Abstract

In unsupervised domain adaptive (UDA) semantic segmentation, the distillation based methods are currently dominant in performance. However, the distillation technique requires complicate multi-stage process and many training tricks. In this paper, we propose a simple yet effective method that can achieve competitive performance to the advanced distillation methods. Our core idea is to fully explore the target-domain information from the views of boundaries and features. First, we propose a novel mix-up strategy to generate high-quality target-domain boundaries with ground-truth labels. Different from the source-domain boundaries in previous works, we select the high-confidence target-domain areas and then paste them to the source-domain images. Such a strategy can generate the object boundaries in target domain (edge of target-domain object areas) with the correct labels. Consequently, the boundary information of target domain can be effectively captured by learning on the mixed-up samples. Second, we design a multi-level contrastive loss to improve the representation of target-domain data, including pixel-level and prototype-level contrastive learning. By combining two proposed methods, more discriminative features can be extracted and hard object boundaries can be better addressed for the target domain. The experimental results on two commonly adopted benchmarks (\textit{i.e.}, GTA5 $\rightarrow$ Cityscapes and SYNTHIA $\rightarrow$ Cityscapes) show that our method achieves competitive performance to complicated distillation methods. Notably, for the SYNTHIA$\rightarrow$ Cityscapes scenario, our method achieves the state-of-the-art performance with $57.8\%$ mIoU and $64.6\%$ mIoU on 16 classes and 13 classes. Code is available at https://github.com/ljjcoder/EHTDI.

Results

TaskDatasetMetricValueModel
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU62EHTDI*
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU62EHTDI*
Image-to-Image TranslationGTAV-to-Cityscapes LabelsmIoU58.8EHTDI(ResNet-101)
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (13 classes)69.2EHTDI*
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (16 classes)61.3EHTDI*
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (13 classes)64.6EHTDI
Image-to-Image TranslationSYNTHIA-to-CityscapesMIoU (16 classes)57.8EHTDI
Domain AdaptationGTA5 to CityscapesmIoU62EHTDI*
Image GenerationGTAV-to-Cityscapes LabelsmIoU62EHTDI*
Image GenerationGTAV-to-Cityscapes LabelsmIoU62EHTDI*
Image GenerationGTAV-to-Cityscapes LabelsmIoU58.8EHTDI(ResNet-101)
Image GenerationSYNTHIA-to-CityscapesMIoU (13 classes)69.2EHTDI*
Image GenerationSYNTHIA-to-CityscapesMIoU (16 classes)61.3EHTDI*
Image GenerationSYNTHIA-to-CityscapesMIoU (13 classes)64.6EHTDI
Image GenerationSYNTHIA-to-CityscapesMIoU (16 classes)57.8EHTDI
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU62EHTDI*
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU62EHTDI*
1 Image, 2*2 StitchingGTAV-to-Cityscapes LabelsmIoU58.8EHTDI(ResNet-101)
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (13 classes)69.2EHTDI*
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (16 classes)61.3EHTDI*
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (13 classes)64.6EHTDI
1 Image, 2*2 StitchingSYNTHIA-to-CityscapesMIoU (16 classes)57.8EHTDI

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17Unified Medical Image Segmentation with State Space Modeling Snake2025-07-17A Privacy-Preserving Semantic-Segmentation Method Using Domain-Adaptation Technique2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17