TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Two at Once: Enhancing Learning and Generalization Capacit...

Two at Once: Enhancing Learning and Generalization Capacities via IBN-Net

Xingang Pan, Ping Luo, Jianping Shi, Xiaoou Tang

2018-07-25ECCV 2018 9All-day Semantic SegmentationDomain GeneralizationRobust Object DetectionVocal Bursts Valence PredictionDomain Adaptation
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Convolutional neural networks (CNNs) have achieved great successes in many computer vision problems. Unlike existing works that designed CNN architectures to improve performance on a single task of a single domain and not generalizable, we present IBN-Net, a novel convolutional architecture, which remarkably enhances a CNN's modeling ability on one domain (e.g. Cityscapes) as well as its generalization capacity on another domain (e.g. GTA5) without finetuning. IBN-Net carefully integrates Instance Normalization (IN) and Batch Normalization (BN) as building blocks, and can be wrapped into many advanced deep networks to improve their performances. This work has three key contributions. (1) By delving into IN and BN, we disclose that IN learns features that are invariant to appearance changes, such as colors, styles, and virtuality/reality, while BN is essential for preserving content related information. (2) IBN-Net can be applied to many advanced deep architectures, such as DenseNet, ResNet, ResNeXt, and SENet, and consistently improve their performance without increasing computational cost. (3) When applying the trained networks to new domains, e.g. from GTA5 to Cityscapes, IBN-Net achieves comparable improvements as domain adaptation methods, even without using data from the target domain. With IBN-Net, we won the 1st place on the WAD 2018 Challenge Drivable Area track, with an mIoU of 86.18%.

Results

TaskDatasetMetricValueModel
Domain AdaptationGTA-to-Avg(Cityscapes,BDD,Mapillary)mIoU34.63IBN
Object DetectionDWDmPC [AP50]25.5IBN-Net
3DDWDmPC [AP50]25.5IBN-Net
2D Semantic SegmentationAll-day CityScapesmIoU64.5IB-Net
2D ClassificationDWDmPC [AP50]25.5IBN-Net
2D Object DetectionDWDmPC [AP50]25.5IBN-Net
Domain GeneralizationGTA-to-Avg(Cityscapes,BDD,Mapillary)mIoU34.63IBN
16kDWDmPC [AP50]25.5IBN-Net

Related Papers

Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17A Privacy-Preserving Semantic-Segmentation Method Using Domain-Adaptation Technique2025-07-17InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing2025-07-16Domain Borders Are There to Be Crossed With Federated Few-Shot Adaptation2025-07-14From Physics to Foundation Models: A Review of AI-Driven Quantitative Remote Sensing Inversion2025-07-11An Offline Mobile Conversational Agent for Mental Health Support: Learning from Emotional Dialogues and Psychological Texts with Student-Centered Evaluation2025-07-11