TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/End-to-End Learnable Geometric Vision by Backpropagating P...

End-to-End Learnable Geometric Vision by Backpropagating PnP Optimization

Bo Chen, Alvaro Parra, Jiewei Cao, Nan Li, Tat-Jun Chin

2019-09-13CVPR 2020 66D Pose Estimation using RGB6D Pose Estimation
PaperPDFCode(official)Code

Abstract

Deep networks excel in learning patterns from large amounts of data. On the other hand, many geometric vision tasks are specified as optimization problems. To seamlessly combine deep learning and geometric vision, it is vital to perform learning and geometric optimization end-to-end. Towards this aim, we present BPnP, a novel network module that backpropagates gradients through a Perspective-n-Points (PnP) solver to guide parameter updates of a neural network. Based on implicit differentiation, we show that the gradients of a "self-contained" PnP solver can be derived accurately and efficiently, as if the optimizer block were a differentiable function. We validate BPnP by incorporating it in a deep model that can learn camera intrinsics, camera extrinsics (poses) and 3D structure from training datasets. Further, we develop an end-to-end trainable pipeline for object pose estimation, which achieves greater accuracy by combining feature-based heatmap losses with 2D-3D reprojection errors. Since our approach can be extended to other optimization problems, our work helps to pave the way to perform learnable geometric vision in a principled manner. Our PyTorch implementation of BPnP is available on http://github.com/BoChenYS/BPnP.

Results

TaskDatasetMetricValueModel
Pose EstimationLineMODMean ADD93.3HRNet (DSNT + BPnP)
3DLineMODMean ADD93.3HRNet (DSNT + BPnP)
1 Image, 2*2 StitchiLineMODMean ADD93.3HRNet (DSNT + BPnP)

Related Papers

SenseShift6D: Multimodal RGB-D Benchmarking for Robust 6D Pose Estimation across Environment and Sensor Variations2025-07-08RAG-6DPose: Retrieval-Augmented 6D Pose Estimation via Leveraging CAD as Knowledge Base2025-06-23Monocular One-Shot Metric-Depth Alignment for RGB-Based Robot Grasping2025-06-20RGBTrack: Fast, Robust Depth-Free 6D Pose Estimation and Tracking2025-06-20Accurate and efficient zero-shot 6D pose estimation with frozen foundation models2025-06-11CHIP: A multi-sensor dataset for 6D pose estimation of chairs in industrial settings2025-06-11UA-Pose: Uncertainty-Aware 6D Object Pose Estimation and Online Object Completion with Partial References2025-06-09You Only Estimate Once: Unified, One-stage, Real-Time Category-level Articulated Object 6D Pose Estimation for Robotic Grasping2025-06-06