TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Cascaded Context Pyramid for Full-Resolution 3D Semantic S...

Cascaded Context Pyramid for Full-Resolution 3D Semantic Scene Completion

Pingping Zhang, Wei Liu, Yinjie Lei, Huchuan Lu, Xiaoyun Yang

2019-08-01ICCV 2019 103D Semantic Scene Completion
PaperPDF

Abstract

Semantic Scene Completion (SSC) aims to simultaneously predict the volumetric occupancy and semantic category of a 3D scene. It helps intelligent devices to understand and interact with the surrounding scenes. Due to the high-memory requirement, current methods only produce low-resolution completion predictions, and generally lose the object details. Furthermore, they also ignore the multi-scale spatial contexts, which play a vital role for the 3D inference. To address these issues, in this work we propose a novel deep learning framework, named Cascaded Context Pyramid Network (CCPNet), to jointly infer the occupancy and semantic labels of a volumetric 3D scene from a single depth image. The proposed CCPNet improves the labeling coherence with a cascaded context pyramid. Meanwhile, based on the low-level features, it progressively restores the fine-structures of objects with Guided Residual Refinement (GRR) modules. Our proposed framework has three outstanding advantages: (1) it explicitly models the 3D spatial context for performance improvement; (2) full-resolution 3D volumes are produced with structure-preserving details; (3) light-weight models with low-memory requirements are captured with a good extensibility. Extensive experiments demonstrate that in spite of taking a single-view depth map, our proposed framework can generate high-quality SSC results, and outperforms state-of-the-art approaches on both the synthetic SUNCG and real NYU datasets.

Results

TaskDatasetMetricValueModel
3D ReconstructionNYUv2mIoU41.3CCPNet (SUNCG pretraining)
3DNYUv2mIoU41.3CCPNet (SUNCG pretraining)
3D Semantic Scene CompletionNYUv2mIoU41.3CCPNet (SUNCG pretraining)

Related Papers

Disentangling Instance and Scene Contexts for 3D Semantic Scene Completion2025-07-11Camera-Only 3D Panoptic Scene Completion for Autonomous Driving through Differentiable Object Shapes2025-05-14SGFormer: Satellite-Ground Fusion for 3D Semantic Scene Completion2025-03-21VLScene: Vision-Language Guidance Distillation for Camera-Based 3D Semantic Scene Completion2025-03-08Vision-based 3D Semantic Scene Completion via Capture Dynamic Representations2025-03-08Learning Temporal 3D Semantic Scene Completion via Optical Flow Guidance2025-02-20Skip Mamba Diffusion for Monocular 3D Semantic Scene Completion2025-01-13SOAP: Vision-Centric 3D Semantic Scene Completion with Scene-Adaptive Decoder and Occluded Region-Aware View Projection2025-01-01