TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Computer Vision/Panoptic Segmentation/ADE20K val

Panoptic Segmentation on ADE20K val

Metric: mIoU (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕mIoU▼Extra DataPaperDate↕Code
1OneFormer (InternImage-H, emb_dim=256, single-scale, 896x896)60.4NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
2X-Decoder (Davit-d5, Deform, single-scale, 1280x1280)59.1YesGeneralized Decoding for Pixel, Image, and Langu...2022-12-21Code
3OneFormer (DiNAT-L, single-scale, 1280x1280, COCO-Pretrain)58.9YesOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
4OneFormer (DiNAT-L, single-scale, 1280x1280)58.3NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
5OneFormer (DiNAT-L, single-scale, 640x640)58.3NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
6X-Decoder (L)58.1YesGeneralized Decoding for Pixel, Image, and Langu...2022-12-21Code
7OneFormer (ConvNeXt-XL, single-scale, 640x640)57.4NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
8OneFormer (Swin-L, single-scale, 1280x1280)57NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
9OneFormer (Swin-L, single-scale, 640x640)57NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
10OneFormer (ConvNeXt-L, single-scale, 640x640)56.6NoOneFormer: One Transformer to Rule Universal Ima...2022-11-10Code
11DiNAT-L (Mask2Former, 640x640)56.3NoDilated Neighborhood Attention Transformer2022-09-29Code
12Mask2Former (Swin-L + FAPN, 640x640)55.4NoMasked-attention Mask Transformer for Universal ...2021-12-02Code
13kMaX-DeepLab (ConvNeXt-L, single-scale, 1281x1281)55.2NokMaX-DeepLab: k-means Mask Transformer2022-07-08Code
14kMaX-DeepLab (ConvNeXt-L, single-scale, 641x641)54.8NokMaX-DeepLab: k-means Mask Transformer2022-07-08Code
15Mask2Former (Swin-L)54.5NoMasked-attention Mask Transformer for Universal ...2021-12-02Code
16Panoptic-DeepLab (SwideRNet)50NoMasked-attention Mask Transformer for Universal ...2021-12-02Code
17Mask2Former (ResNet-50, 640x640)46.1NoMasked-attention Mask Transformer for Universal ...2021-12-02Code
18kMaX-DeepLab (ResNet50, single-scale, 1281x1281)45.3NokMaX-DeepLab: k-means Mask Transformer2022-07-08Code
19kMaX-DeepLab (ResNet50, single-scale, 641x641)45NokMaX-DeepLab: k-means Mask Transformer2022-07-08Code