Weikai Tan, Nannan Qin, Lingfei Ma, Ying Li, Jing Du, Guorong Cai, Ke Yang, Jonathan Li
Semantic segmentation of large-scale outdoor point clouds is essential for urban scene understanding in various applications, especially autonomous driving and urban high-definition (HD) mapping. With rapid developments of mobile laser scanning (MLS) systems, massive point clouds are available for scene understanding, but publicly accessible large-scale labeled datasets, which are essential for developing learning-based methods, are still limited. This paper introduces Toronto-3D, a large-scale urban outdoor point cloud dataset acquired by a MLS system in Toronto, Canada for semantic segmentation. This dataset covers approximately 1 km of point clouds and consists of about 78.3 million points with 8 labeled object classes. Baseline experiments for semantic segmentation were conducted and the results confirmed the capability of this dataset to train deep learning models effectively. Toronto-3D is released to encourage new research, and the labels will be improved and updated with feedback from the research community.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Semantic Segmentation | Toronto-3D | OA | 91.71 | KPFCNN |
| Semantic Segmentation | Toronto-3D | mIoU | 60.3 | KPFCNN |
| Semantic Segmentation | Toronto-3D | OA | 91.64 | TGNet |
| Semantic Segmentation | Toronto-3D | mIoU | 58.34 | TGNet |
| Semantic Segmentation | Toronto-3D | OA | 91.53 | MS-PCNN |
| Semantic Segmentation | Toronto-3D | mIoU | 58.01 | MS-PCNN |
| Semantic Segmentation | Toronto-3D | OA | 91.21 | PointNet++ |
| Semantic Segmentation | Toronto-3D | mIoU | 56.55 | PointNet++ |
| Semantic Segmentation | Toronto-3D | OA | 89 | DGCNN |
| Semantic Segmentation | Toronto-3D | mIoU | 49.6 | DGCNN |
| 3D Semantic Segmentation | Toronto-3D | OA | 91.71 | KPFCNN |
| 3D Semantic Segmentation | Toronto-3D | mIoU | 60.3 | KPFCNN |
| 3D Semantic Segmentation | Toronto-3D | OA | 91.64 | TGNet |
| 3D Semantic Segmentation | Toronto-3D | mIoU | 58.34 | TGNet |
| 3D Semantic Segmentation | Toronto-3D | OA | 91.53 | MS-PCNN |
| 3D Semantic Segmentation | Toronto-3D | mIoU | 58.01 | MS-PCNN |
| 3D Semantic Segmentation | Toronto-3D | OA | 91.21 | PointNet++ |
| 3D Semantic Segmentation | Toronto-3D | mIoU | 56.55 | PointNet++ |
| 3D Semantic Segmentation | Toronto-3D | OA | 89 | DGCNN |
| 3D Semantic Segmentation | Toronto-3D | mIoU | 49.6 | DGCNN |
| 10-shot image generation | Toronto-3D | OA | 91.71 | KPFCNN |
| 10-shot image generation | Toronto-3D | mIoU | 60.3 | KPFCNN |
| 10-shot image generation | Toronto-3D | OA | 91.64 | TGNet |
| 10-shot image generation | Toronto-3D | mIoU | 58.34 | TGNet |
| 10-shot image generation | Toronto-3D | OA | 91.53 | MS-PCNN |
| 10-shot image generation | Toronto-3D | mIoU | 58.01 | MS-PCNN |
| 10-shot image generation | Toronto-3D | OA | 91.21 | PointNet++ |
| 10-shot image generation | Toronto-3D | mIoU | 56.55 | PointNet++ |
| 10-shot image generation | Toronto-3D | OA | 89 | DGCNN |
| 10-shot image generation | Toronto-3D | mIoU | 49.6 | DGCNN |