TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Task-aligned Part-aware Panoptic Segmentation through Join...

Task-aligned Part-aware Panoptic Segmentation through Joint Object-Part Representations

Daan de Geus, Gijs Dubbelman

2024-06-14CVPR 2024 1Panoptic SegmentationPart-aware Panoptic Segmentation
PaperPDFCode(official)

Abstract

Part-aware panoptic segmentation (PPS) requires (a) that each foreground object and background region in an image is segmented and classified, and (b) that all parts within foreground objects are segmented, classified and linked to their parent object. Existing methods approach PPS by separately conducting object-level and part-level segmentation. However, their part-level predictions are not linked to individual parent objects. Therefore, their learning objective is not aligned with the PPS task objective, which harms the PPS performance. To solve this, and make more accurate PPS predictions, we propose Task-Aligned Part-aware Panoptic Segmentation (TAPPS). This method uses a set of shared queries to jointly predict (a) object-level segments, and (b) the part-level segments within those same objects. As a result, TAPPS learns to predict part-level segments that are linked to individual parent objects, aligning the learning objective with the task objective, and allowing TAPPS to leverage joint object-part representations. With experiments, we show that TAPPS considerably outperforms methods that predict objects and parts separately, and achieves new state-of-the-art PPS results.

Results

TaskDatasetMetricValueModel
Part-aware Panoptic SegmentationCityscapes Panoptic PartsPartPQ64.8TAPPS (Swin-B, COCO pre-training)
Part-aware Panoptic SegmentationPascal Panoptic PartsPartPQ60.4TAPPS (Swin-B, COCO pre-training)

Related Papers

DEARLi: Decoupled Enhancement of Recognition and Localization for Semi-supervised Panoptic Segmentation2025-07-14OpenWorldSAM: Extending SAM2 for Universal Image Segmentation with Language Prompts2025-07-07HieraSurg: Hierarchy-Aware Diffusion Model for Surgical Video Generation2025-06-26PanSt3R: Multi-view Consistent Panoptic Segmentation2025-06-26Open-Set LiDAR Panoptic Segmentation Guided by Uncertainty-Aware Learning2025-06-16A Comprehensive Survey on Video Scene Parsing:Advances, Challenges, and Prospects2025-06-16The Missing Point in Vision Transformers for Universal Image Segmentation2025-05-26How Do Images Align and Complement LiDAR? Towards a Harmonized Multi-modal 3D Panoptic Segmentation2025-05-25