TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Anisotropic Convolutional Networks for 3D Semantic Scene C...

Anisotropic Convolutional Networks for 3D Semantic Scene Completion

Jie Li, Kai Han, Peng Wang, Yu Liu, Xia Yuan

2020-04-05CVPR 2020 63D Semantic Scene Completion from a single RGB image3D Semantic Scene Completion
PaperPDFCode(official)

Abstract

As a voxel-wise labeling task, semantic scene completion (SSC) tries to simultaneously infer the occupancy and semantic labels for a scene from a single depth and/or RGB image. The key challenge for SSC is how to effectively take advantage of the 3D context to model various objects or stuffs with severe variations in shapes, layouts and visibility. To handle such variations, we propose a novel module called anisotropic convolution, which properties with flexibility and power impossible for the competing methods such as standard 3D convolution and some of its variations. In contrast to the standard 3D convolution that is limited to a fixed 3D receptive field, our module is capable of modeling the dimensional anisotropy voxel-wisely. The basic idea is to enable anisotropic 3D receptive field by decomposing a 3D convolution into three consecutive 1D convolutions, and the kernel size for each such 1D convolution is adaptively determined on the fly. By stacking multiple such anisotropic convolution modules, the voxel-wise modeling capability can be further enhanced while maintaining a controllable amount of model parameters. Extensive experiments on two SSC benchmarks, NYU-Depth-v2 and NYUCAD, show the superior performance of the proposed method. Our code is available at https://waterljwant.github.io/SSC/

Results

TaskDatasetMetricValueModel
ReconstructionNYUv2mIoU18.15AICNet (rgb input - reported in MonoScene paper)
3D ReconstructionNYUv2mIoU33.3AIC-Net
3D ReconstructionNYUv2mIoU18.15AICNet (rgb input - reported in MonoScene paper)
3DNYUv2mIoU33.3AIC-Net
3DNYUv2mIoU18.15AICNet (rgb input - reported in MonoScene paper)
3D Semantic Scene CompletionNYUv2mIoU33.3AIC-Net
3D Semantic Scene CompletionNYUv2mIoU18.15AICNet (rgb input - reported in MonoScene paper)
3D Scene ReconstructionNYUv2mIoU18.15AICNet (rgb input - reported in MonoScene paper)
Single-View 3D ReconstructionNYUv2mIoU18.15AICNet (rgb input - reported in MonoScene paper)

Related Papers

Disentangling Instance and Scene Contexts for 3D Semantic Scene Completion2025-07-11Camera-Only 3D Panoptic Scene Completion for Autonomous Driving through Differentiable Object Shapes2025-05-14SGFormer: Satellite-Ground Fusion for 3D Semantic Scene Completion2025-03-21VLScene: Vision-Language Guidance Distillation for Camera-Based 3D Semantic Scene Completion2025-03-08Vision-based 3D Semantic Scene Completion via Capture Dynamic Representations2025-03-08Learning Temporal 3D Semantic Scene Completion via Optical Flow Guidance2025-02-20Skip Mamba Diffusion for Monocular 3D Semantic Scene Completion2025-01-13SOAP: Vision-Centric 3D Semantic Scene Completion with Scene-Adaptive Decoder and Occluded Region-Aware View Projection2025-01-01