TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Domain Generalization Using Large Pretrained Models with M...

Domain Generalization Using Large Pretrained Models with Mixture-of-Adapters

Gyuseong Lee, Wooseok Jang, Jinhyeon Kim, Jaewoo Jung, Seungryong Kim

2023-10-17Domain Generalizationparameter-efficient fine-tuning
PaperPDFCode(official)

Abstract

Learning robust vision models that perform well in out-of-distribution (OOD) situations is an important task for model deployment in real-world settings. Despite extensive research in this field, many proposed methods have only shown minor performance improvements compared to the simplest empirical risk minimization (ERM) approach, which was evaluated on a benchmark with a limited hyperparameter search space. Our focus in this study is on leveraging the knowledge of large pretrained models to improve handling of OOD scenarios and tackle domain generalization problems. However, prior research has revealed that naively fine-tuning a large pretrained model can impair OOD robustness. Thus, we employ parameter-efficient fine-tuning (PEFT) techniques to effectively preserve OOD robustness while working with large models. Our extensive experiments and analysis confirm that the most effective approaches involve ensembling diverse models and increasing the scale of pretraining. As a result, we achieve state-of-the-art performance in domain generalization tasks. Our code and project page are available at: https://cvlab-kaist.github.io/MoA

Results

TaskDatasetMetricValueModel
Domain AdaptationPACSAverage Accuracy97.4MoA (OpenCLIP, ViT-B/16)
Domain AdaptationOffice-HomeAverage Accuracy90.6MoA (OpenCLIP, ViT-B/16)
Domain AdaptationDomainNetAverage Accuracy62.7MoA (OpenCLIP, ViT-B/16)
Domain AdaptationVLCSAverage Accuracy83.1MoA (OpenCLIP, ViT-B/16)
Domain AdaptationTerraIncognitaAverage Accuracy52.8MoA (OpenCLIP, ViT-B/16)
Domain GeneralizationPACSAverage Accuracy97.4MoA (OpenCLIP, ViT-B/16)
Domain GeneralizationOffice-HomeAverage Accuracy90.6MoA (OpenCLIP, ViT-B/16)
Domain GeneralizationDomainNetAverage Accuracy62.7MoA (OpenCLIP, ViT-B/16)
Domain GeneralizationVLCSAverage Accuracy83.1MoA (OpenCLIP, ViT-B/16)
Domain GeneralizationTerraIncognitaAverage Accuracy52.8MoA (OpenCLIP, ViT-B/16)

Related Papers

Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing2025-07-16From Physics to Foundation Models: A Review of AI-Driven Quantitative Remote Sensing Inversion2025-07-11Feed-Forward SceneDINO for Unsupervised Semantic Scene Completion2025-07-08Prompt-Free Conditional Diffusion for Multi-object Image Augmentation2025-07-08