TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Progressive Random Convolutions for Single Domain Generali...

Progressive Random Convolutions for Single Domain Generalization

Seokeon Choi, Debasmit Das, Sungha Choi, Seunghan Yang, Hyunsin Park, Sungrack Yun

2023-04-02CVPR 2023 1Photo to Rest GeneralizationImage to sketch recognitionImage AugmentationDomain GeneralizationSingle-Source Domain Generalization
PaperPDF

Abstract

Single domain generalization aims to train a generalizable model with only one source domain to perform well on arbitrary unseen target domains. Image augmentation based on Random Convolutions (RandConv), consisting of one convolution layer randomly initialized for each mini-batch, enables the model to learn generalizable visual representations by distorting local textures despite its simple and lightweight structure. However, RandConv has structural limitations in that the generated image easily loses semantics as the kernel size increases, and lacks the inherent diversity of a single convolution operation. To solve the problem, we propose a Progressive Random Convolution (Pro-RandConv) method that recursively stacks random convolution layers with a small kernel size instead of increasing the kernel size. This progressive approach can not only mitigate semantic distortions by reducing the influence of pixels away from the center in the theoretical receptive field, but also create more effective virtual domains by gradually increasing the style diversity. In addition, we develop a basic random convolution layer into a random convolution block including deformable offsets and affine transformation to support texture and contrast diversification, both of which are also randomly initialized. Without complex generators or adversarial learning, we demonstrate that our simple yet effective augmentation strategy outperforms state-of-the-art methods on single domain generalization benchmarks.

Results

TaskDatasetMetricValueModel
Domain AdaptationPACSAccuracy68.88ProRandConv (ResNet18)
Domain AdaptationDigits-fiveAccuracy81.35ProRandConv (LeNet)
Domain AdaptationPACSAccuracy62.89ProRandConv (ResNet18)
Domain GeneralizationPACSAccuracy68.88ProRandConv (ResNet18)
Domain GeneralizationDigits-fiveAccuracy81.35ProRandConv (LeNet)
Domain GeneralizationPACSAccuracy62.89ProRandConv (ResNet18)
Single-Source Domain GeneralizationPACSAccuracy68.88ProRandConv (ResNet18)
Single-Source Domain GeneralizationDigits-fiveAccuracy81.35ProRandConv (LeNet)
Single-Source Domain GeneralizationPACSAccuracy62.89ProRandConv (ResNet18)

Related Papers

Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17InstructFLIP: Exploring Unified Vision-Language Model for Face Anti-spoofing2025-07-16From Physics to Foundation Models: A Review of AI-Driven Quantitative Remote Sensing Inversion2025-07-11Prompt-Free Conditional Diffusion for Multi-object Image Augmentation2025-07-08Feed-Forward SceneDINO for Unsupervised Semantic Scene Completion2025-07-08Integrated Structural Prompt Learning for Vision-Language Models2025-07-08