TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Medical/Semantic Segmentation/ADE20K val

Semantic Segmentation on ADE20K val

Metric: AP (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕AP▼Extra DataPaperDate↕Code
1OneFormer (InternImage-H, emb_dim=256, single-scale, 896x896)40.2NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
2X-Decoder (Davit-d5, Deform, single-scale, 1280x1280)38.7YesGeneralized Decoding for Pixel, Image, and Langu...2022-12-21Code
3OneFormer (Swin-L, single-scale, 1280x1280)37.8NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
4OneFormer (DiNAT-L, single-scale, 1280x1280)37.1NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
5OneFormer (ConvNeXt-XL, single-scale, 640x640)36.3NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
6OneFormer (ConvNeXt-L, single-scale, 640x640)36.2NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
7OneFormer (DiNAT-L, single-scale, 640x640)36NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
8OneFormer (Swin-L, single-scale, 640x640)35.9NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
9X-Decoder (L)35.8YesGeneralized Decoding for Pixel, Image, and Langu...2022-12-21Code
10DiNAT-L (Mask2Former, 640x640)35NoDilated Neighborhood Attention Transformer2022-09-29Code
11Mask2Former (Swin-L)34.2NoMasked-attention Mask Transformer for Universal ...2021-12-02Code
12Mask2Former (Swin-L + FAPN, 640x640)33.2NoMasked-attention Mask Transformer for Universal ...2021-12-02Code
13Mask2Former (ResNet-50, 640x640)26.5NoMasked-attention Mask Transformer for Universal ...2021-12-02Code