TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Wave-SAN: Wavelet based Style Augmentation Network for Cro...

Wave-SAN: Wavelet based Style Augmentation Network for Cross-Domain Few-Shot Learning

Yuqian Fu, Yu Xie, Yanwei Fu, Jingjing Chen, Yu-Gang Jiang

2022-03-15Few-Shot LearningSelf-Supervised LearningCross-Domain Few-Shotcross-domain few-shot learning
PaperPDFCode(official)

Abstract

Previous few-shot learning (FSL) works mostly are limited to natural images of general concepts and categories. These works assume very high visual similarity between the source and target classes. In contrast, the recently proposed cross-domain few-shot learning (CD-FSL) aims at transferring knowledge from general nature images of many labeled examples to novel domain-specific target categories of only a few labeled examples. The key challenge of CD-FSL lies in the huge data shift between source and target domains, which is typically in the form of totally different visual styles. This makes it very nontrivial to directly extend the classical FSL methods to address the CD-FSL task. To this end, this paper studies the problem of CD-FSL by spanning the style distributions of the source dataset. Particularly, wavelet transform is introduced to enable the decomposition of visual representations into low-frequency components such as shape and style and high-frequency components e.g., texture. To make our model robust to visual styles, the source images are augmented by swapping the styles of their low-frequency components with each other. We propose a novel Style Augmentation (StyleAug) module to implement this idea. Furthermore, we present a Self-Supervised Learning (SSL) module to ensure the predictions of style-augmented images are semantically similar to the unchanged ones. This avoids the potential semantic drift problem in exchanging the styles. Extensive experiments on two CD-FSL benchmarks show the effectiveness of our method. Our codes and models will be released.

Results

TaskDatasetMetricValueModel
Few-Shot LearningChestX5 shot25.63wave-SAN
Few-Shot LearningPlantae5 shot57.72wave-SAN
Few-Shot Learningcars5 shot46.11wave-SAN
Few-Shot LearningEuroSAT5 shot85.22wave-SAN
Few-Shot LearningCUB5 shot70.31wave-SAN
Few-Shot LearningISIC20185 shot44.93wave-SAN
Few-Shot LearningPlaces5 shot76.88wave-SAN
Few-Shot LearningCropDisease5 shot89.7wave-SAN
Meta-LearningChestX5 shot25.63wave-SAN
Meta-LearningPlantae5 shot57.72wave-SAN
Meta-Learningcars5 shot46.11wave-SAN
Meta-LearningEuroSAT5 shot85.22wave-SAN
Meta-LearningCUB5 shot70.31wave-SAN
Meta-LearningISIC20185 shot44.93wave-SAN
Meta-LearningPlaces5 shot76.88wave-SAN
Meta-LearningCropDisease5 shot89.7wave-SAN
Cross-Domain Few-ShotChestX5 shot25.63wave-SAN
Cross-Domain Few-ShotPlantae5 shot57.72wave-SAN
Cross-Domain Few-Shotcars5 shot46.11wave-SAN
Cross-Domain Few-ShotEuroSAT5 shot85.22wave-SAN
Cross-Domain Few-ShotCUB5 shot70.31wave-SAN
Cross-Domain Few-ShotISIC20185 shot44.93wave-SAN
Cross-Domain Few-ShotPlaces5 shot76.88wave-SAN
Cross-Domain Few-ShotCropDisease5 shot89.7wave-SAN

Related Papers

GLAD: Generalizable Tuning for Vision-Language Models2025-07-17A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17Self-supervised Learning on Camera Trap Footage Yields a Strong Universal Face Embedder2025-07-14Doodle Your Keypoints: Sketch-Based Few-Shot Keypoint Detection2025-07-10An Enhanced Privacy-preserving Federated Few-shot Learning Framework for Respiratory Disease Diagnosis2025-07-10Few-Shot Learning by Explicit Physics Integration: An Application to Groundwater Heat Transport2025-07-08Speech Quality Assessment Model Based on Mixture of Experts: System-Level Performance Enhancement and Utterance-Level Challenge Analysis2025-07-08ViRefSAM: Visual Reference-Guided Segment Anything Model for Remote Sensing Segmentation2025-07-03