TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Point Cloud Mamba: Point Cloud Learning via State Space Mo...

Point Cloud Mamba: Point Cloud Learning via State Space Model

Tao Zhang, Haobo Yuan, Lu Qi, Jiangning Zhang, Qianyu Zhou, Shunping Ji, Shuicheng Yan, Xiangtai Li

2024-03-01Supervised Only 3D Point Cloud Classification
PaperPDFCode(official)Code(official)

Abstract

Recently, state space models have exhibited strong global modeling capabilities and linear computational complexity in contrast to transformers. This research focuses on applying such architecture to more efficiently and effectively model point cloud data globally with linear computational complexity. In particular, for the first time, we demonstrate that Mamba-based point cloud methods can outperform previous methods based on transformer or multi-layer perceptrons (MLPs). To enable Mamba to process 3-D point cloud data more effectively, we propose a novel Consistent Traverse Serialization method to convert point clouds into 1-D point sequences while ensuring that neighboring points in the sequence are also spatially adjacent. Consistent Traverse Serialization yields six variants by permuting the order of \textit{x}, \textit{y}, and \textit{z} coordinates, and the synergistic use of these variants aids Mamba in comprehensively observing point cloud data. Furthermore, to assist Mamba in handling point sequences with different orders more effectively, we introduce point prompts to inform Mamba of the sequence's arrangement rules. Finally, we propose positional encoding based on spatial coordinate mapping to inject positional information into point cloud sequences more effectively. Point Cloud Mamba surpasses the state-of-the-art (SOTA) point-based method PointNeXt and achieves new SOTA performance on the ScanObjectNN, ModelNet40, ShapeNetPart, and S3DIS datasets. It is worth mentioning that when using a more powerful local feature extraction module, our PCM achieves 79.6 mIoU on S3DIS, significantly surpassing the previous SOTA models, DeLA and PTv3, by 5.5 mIoU and 4.9 mIoU, respectively.

Results

TaskDatasetMetricValueModel
Shape Representation Of 3D Point CloudsScanObjectNNGFLOPs45PCM
Shape Representation Of 3D Point CloudsScanObjectNNNumber of params (M)34.2PCM
Shape Representation Of 3D Point CloudsScanObjectNNOverall Accuracy (PB_T50_RS)88.1PCM
3D Point Cloud ClassificationScanObjectNNGFLOPs45PCM
3D Point Cloud ClassificationScanObjectNNNumber of params (M)34.2PCM
3D Point Cloud ClassificationScanObjectNNOverall Accuracy (PB_T50_RS)88.1PCM
3D Point Cloud ReconstructionScanObjectNNGFLOPs45PCM
3D Point Cloud ReconstructionScanObjectNNNumber of params (M)34.2PCM
3D Point Cloud ReconstructionScanObjectNNOverall Accuracy (PB_T50_RS)88.1PCM

Related Papers

Mamba3D: Enhancing Local Features for 3D Point Cloud Analysis via State Space Model2024-04-23Decoupled Local Aggregation for Point Cloud Learning2023-08-31Self-positioning Point-based Transformer for Point Cloud Understanding2023-03-29Parameter is Not All You Need: Starting from Non-Parametric Networks for 3D Point Cloud Analysis2023-03-14Point Cloud Classification Using Content-based Transformer via Clustering in Feature Space2023-03-08PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies2022-06-09Surface Representation for Point Clouds2022-05-11Rethinking Network Design and Local Geometry in Point Cloud: A Simple Residual MLP Framework2022-02-15