TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Audio/10-shot image generation/ADE20K

10-shot image generation on ADE20K

Metric: GFLOPs (512 x 512) (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕GFLOPs (512 x 512)▼Extra DataPaperDate↕Code
1ConvNeXt-XL++3335NoA ConvNet for the 2020s2022-01-10Code
2ConvNeXt-L++2458NoA ConvNet for the 2020s2022-01-10Code
3ConvNeXt-B++1828NoA ConvNet for the 2020s2022-01-10Code
4GC ViT-B1348NoGlobal Context Vision Transformers2022-06-20Code
5MogaNet-L (UperNet)1176NoMogaNet: Multi-order Gated Aggregation Network2022-11-07Code
6ConvNeXt-B1170NoA ConvNet for the 2020s2022-01-10Code
7GC ViT-S1163NoGlobal Context Vision Transformers2022-06-20Code
8NAT-Base1137NoNeighborhood Attention Transformer2022-04-14Code
9MogaNet-B (UperNet)1050NoMogaNet: Multi-order Gated Aggregation Network2022-11-07Code
10ConvNeXt-S1027NoA ConvNet for the 2020s2022-01-10Code
11NAT-Small1010NoNeighborhood Attention Transformer2022-04-14Code
12GC ViT-T947NoGlobal Context Vision Transformers2022-06-20Code
13MogaNet-S (UperNet)946NoMogaNet: Multi-order Gated Aggregation Network2022-11-07Code
14ConvNeXt-T939NoA ConvNet for the 2020s2022-01-10Code
15NAT-Tiny934NoNeighborhood Attention Transformer2022-04-14Code
16NAT-Mini900NoNeighborhood Attention Transformer2022-04-14Code
17EoMT (DINOv2-L, single-scale, 512x512)721NoYour ViT is Secretly an Image Segmentation Model2025-03-24Code
18SegViT-v2 (BEiT-v2-Large)637.9NoSegViTv2: Exploring Efficient and Continual Sema...2023-06-09Code
19MogaNet-S (Semantic FPN)189NoMogaNet: Multi-order Gated Aggregation Network2022-11-07Code
20Light-Ham (VAN-Huge)71.8NoIs Attention Better Than Matrix Decomposition?2021-09-09Code
21HRViT-b3 (SegFormer, SS)67.9NoMulti-Scale High-Resolution Vision Transformer f...2021-11-01Code
22Light-Ham (VAN-Large)55NoIs Attention Better Than Matrix Decomposition?2021-09-09Code
23Light-Ham (VAN-Base)34.4NoIs Attention Better Than Matrix Decomposition?2021-09-09Code
24HRViT-b2 (SegFormer, SS)28NoMulti-Scale High-Resolution Vision Transformer f...2021-11-01Code
25Light-Ham (VAN-Small, D=256)15.8NoIs Attention Better Than Matrix Decomposition?2021-09-09Code
26HRViT-b1 (SegFormer, SS)14.6NoMulti-Scale High-Resolution Vision Transformer f...2021-11-01Code