TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Sparse Mixture-of-Experts are Domain Generalizable Learners

Sparse Mixture-of-Experts are Domain Generalizable Learners

Bo Li, Yifei Shen, Jingkang Yang, Yezhen Wang, Jiawei Ren, Tong Che, Jun Zhang, Ziwei Liu

2022-06-08Domain GeneralizationObject Recognition
PaperPDFCode(official)Code

Abstract

Human visual perception can easily generalize to out-of-distributed visual data, which is far beyond the capability of modern machine learning models. Domain generalization (DG) aims to close this gap, with existing DG methods mainly focusing on the loss function design. In this paper, we propose to explore an orthogonal direction, i.e., the design of the backbone architecture. It is motivated by an empirical finding that transformer-based models trained with empirical risk minimization (ERM) outperform CNN-based models employing state-of-the-art (SOTA) DG algorithms on multiple DG datasets. We develop a formal framework to characterize a network's robustness to distribution shifts by studying its architecture's alignment with the correlations in the dataset. This analysis guides us to propose a novel DG model built upon vision transformers, namely Generalizable Mixture-of-Experts (GMoE). Extensive experiments on DomainBed demonstrate that GMoE trained with ERM outperforms SOTA DG baselines by a large margin. Moreover, GMoE is complementary to existing DG methods and its performance is substantially improved when trained with DG algorithms.

Results

TaskDatasetMetricValueModel
Domain AdaptationPACSAverage Accuracy88.1GMoE-S/16
Domain AdaptationOffice-HomeAverage Accuracy74.2GMoE-S/16
Domain AdaptationDomainNetAverage Accuracy52Hybrid-SF-MoE
Domain AdaptationDomainNetAverage Accuracy48.7GMoE-S/16
Domain AdaptationVLCSAverage Accuracy80.2GMoE-S/16
Domain AdaptationTerraIncognitaAverage Accuracy48.5GMoE-S/16
Domain GeneralizationPACSAverage Accuracy88.1GMoE-S/16
Domain GeneralizationOffice-HomeAverage Accuracy74.2GMoE-S/16
Domain GeneralizationDomainNetAverage Accuracy52Hybrid-SF-MoE
Domain GeneralizationDomainNetAverage Accuracy48.7GMoE-S/16
Domain GeneralizationVLCSAverage Accuracy80.2GMoE-S/16
Domain GeneralizationTerraIncognitaAverage Accuracy48.5GMoE-S/16

Related Papers

Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing2025-07-16From Physics to Foundation Models: A Review of AI-Driven Quantitative Remote Sensing Inversion2025-07-11Feed-Forward SceneDINO for Unsupervised Semantic Scene Completion2025-07-08Prompt-Free Conditional Diffusion for Multi-object Image Augmentation2025-07-08Integrated Structural Prompt Learning for Vision-Language Models2025-07-08