TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Points to Patches: Enabling the Use of Self-Attention for ...

Points to Patches: Enabling the Use of Self-Attention for 3D Shape Recognition

Axel Berg, Magnus Oskarsson, Mark O'Connor

2022-04-08Point Cloud Registration3D Shape Recognition3D Point Cloud Classification3D Feature Matching
PaperPDFCode(official)

Abstract

While the Transformer architecture has become ubiquitous in the machine learning field, its adaptation to 3D shape recognition is non-trivial. Due to its quadratic computational complexity, the self-attention operator quickly becomes inefficient as the set of input points grows larger. Furthermore, we find that the attention mechanism struggles to find useful connections between individual points on a global scale. In order to alleviate these problems, we propose a two-stage Point Transformer-in-Transformer (Point-TnT) approach which combines local and global attention mechanisms, enabling both individual points and patches of points to attend to each other effectively. Experiments on shape classification show that such an approach provides more useful features for downstream tasks than the baseline Transformer, while also being more computationally efficient. In addition, we also extend our method to feature matching for scene reconstruction, showing that it can be used in conjunction with existing scene reconstruction pipelines.

Results

TaskDatasetMetricValueModel
Point Cloud Registration3DMatch BenchmarkFeature Matching Recall96.8DIP + Point-TnT
Shape Representation Of 3D Point CloudsScanObjectNNMean Accuracy81Point-TnT
Shape Representation Of 3D Point CloudsScanObjectNNOverall Accuracy83.5Point-TnT
Shape Representation Of 3D Point CloudsModelNet40Overall Accuracy92.6Point-TnT
3D Point Cloud ClassificationScanObjectNNMean Accuracy81Point-TnT
3D Point Cloud ClassificationScanObjectNNOverall Accuracy83.5Point-TnT
3D Point Cloud ClassificationModelNet40Overall Accuracy92.6Point-TnT
3D Point Cloud Interpolation3DMatch BenchmarkFeature Matching Recall96.8DIP + Point-TnT
3D Point Cloud ReconstructionScanObjectNNMean Accuracy81Point-TnT
3D Point Cloud ReconstructionScanObjectNNOverall Accuracy83.5Point-TnT
3D Point Cloud ReconstructionModelNet40Overall Accuracy92.6Point-TnT

Related Papers

A Multi-Level Similarity Approach for Single-View Object Grasping: Matching, Planning, and Fine-Tuning2025-07-16Simultaneous Localization and Mapping Using Active mmWave Sensing in 5G NR2025-07-07CA-I2P: Channel-Adaptive Registration Network with Global Optimal Selection2025-06-26Asymmetric Dual Self-Distillation for 3D Self-Supervised Representation Learning2025-06-26Correspondence-Free Multiview Point Cloud Registration via Depth-Guided Joint Optimisation2025-06-18MT-PCR: A Hybrid Mamba-Transformer with Spatial Serialization for Hierarchical Point Cloud Registration2025-06-16Robust Filtering -- Novel Statistical Learning and Inference Algorithms with Applications2025-06-13Rectified Point Flow: Generic Point Cloud Pose Estimation2025-06-05