TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Cross-modal feature fusion for robust point cloud registra...

Cross-modal feature fusion for robust point cloud registration with ambiguous geometry

Zhaoyi Wang, Shengyu Huang, Jemil Avers Butt, Yuanzhou Cai, Matej Varga, Andreas Wieser

2025-05-19Point Cloud Registration
PaperPDFCode(official)

Abstract

Point cloud registration has seen significant advancements with the application of deep learning techniques. However, existing approaches often overlook the potential of integrating radiometric information from RGB images. This limitation reduces their effectiveness in aligning point clouds pairs, especially in regions where geometric data alone is insufficient. When used effectively, radiometric information can enhance the registration process by providing context that is missing from purely geometric data. In this paper, we propose CoFF, a novel Cross-modal Feature Fusion method that utilizes both point cloud geometry and RGB images for pairwise point cloud registration. Assuming that the co-registration between point clouds and RGB images is available, CoFF explicitly addresses the challenges where geometric information alone is unclear, such as in regions with symmetric similarity or planar structures, through a two-stage fusion of 3D point cloud features and 2D image features. It incorporates a cross-modal feature fusion module that assigns pixel-wise image features to 3D input point clouds to enhance learned 3D point features, and integrates patch-wise image features with superpoint features to improve the quality of coarse matching. This is followed by a coarse-to-fine matching module that accurately establishes correspondences using the fused features. We extensively evaluate CoFF on four common datasets: 3DMatch, 3DLoMatch, IndoorLRS, and the recently released ScanNet++ datasets. In addition, we assess CoFF on specific subset datasets containing geometrically ambiguous cases. Our experimental results demonstrate that CoFF achieves state-of-the-art registration performance across all benchmarks, including remarkable registration recalls of 95.9% and 81.6% on the widely-used 3DMatch and 3DLoMatch datasets, respectively...(Truncated to fit arXiv abstract length)

Results

TaskDatasetMetricValueModel
Point Cloud Registration3DLoMatch (10-30% overlap)Recall ( correspondence RMSE below 0.2)81.6CoFF
Point Cloud RegistrationScanNet++ (trained on 3DMatch)Recall ( correspondence RMSE below 0.2)78.7CoFF
3D Point Cloud Interpolation3DLoMatch (10-30% overlap)Recall ( correspondence RMSE below 0.2)81.6CoFF
3D Point Cloud InterpolationScanNet++ (trained on 3DMatch)Recall ( correspondence RMSE below 0.2)78.7CoFF

Related Papers

A Multi-Level Similarity Approach for Single-View Object Grasping: Matching, Planning, and Fine-Tuning2025-07-16Simultaneous Localization and Mapping Using Active mmWave Sensing in 5G NR2025-07-07CA-I2P: Channel-Adaptive Registration Network with Global Optimal Selection2025-06-26Correspondence-Free Multiview Point Cloud Registration via Depth-Guided Joint Optimisation2025-06-18MT-PCR: A Hybrid Mamba-Transformer with Spatial Serialization for Hierarchical Point Cloud Registration2025-06-16Robust Filtering -- Novel Statistical Learning and Inference Algorithms with Applications2025-06-13Rectified Point Flow: Generic Point Cloud Pose Estimation2025-06-05MonoMobility: Zero-Shot 3D Mobility Analysis from Monocular Videos2025-05-17