TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CPPF: Towards Robust Category-Level 9D Pose Estimation in ...

CPPF: Towards Robust Category-Level 9D Pose Estimation in the Wild

Yang You, Ruoxi Shi, Weiming Wang, Cewu Lu

2022-03-07CVPR 2022 1Pose Estimation6D Pose Estimation using RGBD
PaperPDFCode(official)

Abstract

In this paper, we tackle the problem of category-level 9D pose estimation in the wild, given a single RGB-D frame. Using supervised data of real-world 9D poses is tedious and erroneous, and also fails to generalize to unseen scenarios. Besides, category-level pose estimation requires a method to be able to generalize to unseen objects at test time, which is also challenging. Drawing inspirations from traditional point pair features (PPFs), in this paper, we design a novel Category-level PPF (CPPF) voting method to achieve accurate, robust and generalizable 9D pose estimation in the wild. To obtain robust pose estimation, we sample numerous point pairs on an object, and for each pair our model predicts necessary SE(3)-invariant voting statistics on object centers, orientations and scales. A novel coarse-to-fine voting algorithm is proposed to eliminate noisy point pair samples and generate final predictions from the population. To get rid of false positives in the orientation voting process, an auxiliary binary disambiguating classification task is introduced for each sampled point pair. In order to detect objects in the wild, we carefully design our sim-to-real pipeline by training on synthetic point clouds only, unless objects have ambiguous poses in geometry. Under this circumstance, color information is leveraged to disambiguate these poses. Results on standard benchmarks show that our method is on par with current state of the arts with real-world training data. Extensive experiments further show that our method is robust to noise and gives promising results under extremely challenging scenarios. Our code is available on https://github.com/qq456cvb/CPPF.

Results

TaskDatasetMetricValueModel
Pose EstimationREAL275mAP 10, 5cm44.9CPPF
Pose EstimationREAL275mAP 15, 5cm50.8CPPF
Pose EstimationREAL275mAP 3DIou@2578.2CPPF
Pose EstimationREAL275mAP 3DIou@5026.4CPPF
Pose EstimationREAL275mAP 5, 5cm16.9CPPF
3DREAL275mAP 10, 5cm44.9CPPF
3DREAL275mAP 15, 5cm50.8CPPF
3DREAL275mAP 3DIou@2578.2CPPF
3DREAL275mAP 3DIou@5026.4CPPF
3DREAL275mAP 5, 5cm16.9CPPF
1 Image, 2*2 StitchiREAL275mAP 10, 5cm44.9CPPF
1 Image, 2*2 StitchiREAL275mAP 15, 5cm50.8CPPF
1 Image, 2*2 StitchiREAL275mAP 3DIou@2578.2CPPF
1 Image, 2*2 StitchiREAL275mAP 3DIou@5026.4CPPF
1 Image, 2*2 StitchiREAL275mAP 5, 5cm16.9CPPF

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16