Weiyue Wang, Ulrich Neumann
Convolutional neural networks (CNN) are limited by the lack of capability to handle geometric information due to the fixed grid kernel structure. The availability of depth data enables progress in RGB-D semantic segmentation with CNNs. State-of-the-art methods either use depth as additional images or process spatial information in 3D volumes or point clouds. These methods suffer from high computation and memory cost. To address these issues, we present Depth-aware CNN by introducing two intuitive, flexible and effective operations: depth-aware convolution and depth-aware average pooling. By leveraging depth similarity between pixels in the process of information propagation, geometry is seamlessly incorporated into CNN. Without introducing any additional parameters, both operators can be easily integrated into existing CNNs. Extensive experiments and ablation studies on challenging RGB-D semantic segmentation benchmarks validate the effectiveness and flexibility of our approach.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Semantic Segmentation | Stanford2D3D - RGBD | Pixel Accuracy | 65.4 | Depth-aware CNN |
| Semantic Segmentation | Stanford2D3D - RGBD | mAcc | 55.5 | Depth-aware CNN |
| Semantic Segmentation | Stanford2D3D - RGBD | mIoU | 39.5 | Depth-aware CNN |
| Semantic Segmentation | MFN Dataset | mIOU | 46.1 | Depth-aware CNN |
| Scene Segmentation | MFN Dataset | mIOU | 46.1 | Depth-aware CNN |
| 2D Object Detection | MFN Dataset | mIOU | 46.1 | Depth-aware CNN |
| 10-shot image generation | Stanford2D3D - RGBD | Pixel Accuracy | 65.4 | Depth-aware CNN |
| 10-shot image generation | Stanford2D3D - RGBD | mAcc | 55.5 | Depth-aware CNN |
| 10-shot image generation | Stanford2D3D - RGBD | mIoU | 39.5 | Depth-aware CNN |
| 10-shot image generation | MFN Dataset | mIOU | 46.1 | Depth-aware CNN |