TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/I^2R-Net: Intra- and Inter-Human Relation Network for Mult...

I^2R-Net: Intra- and Inter-Human Relation Network for Multi-Person Pose Estimation

Yiwei Ding, Wenjin Deng, Yinglin Zheng, PengFei Liu, Meihong Wang, Xuan Cheng, Jianmin Bao, Dong Chen, Ming Zeng

2022-06-22Pose EstimationMulti-Person Pose Estimation
PaperPDFCode(official)

Abstract

In this paper, we present the Intra- and Inter-Human Relation Networks (I^2R-Net) for Multi-Person Pose Estimation. It involves two basic modules. First, the Intra-Human Relation Module operates on a single person and aims to capture Intra-Human dependencies. Second, the Inter-Human Relation Module considers the relation between multiple instances and focuses on capturing Inter-Human interactions. The Inter-Human Relation Module can be designed very lightweight by reducing the resolution of feature map, yet learn useful relation information to significantly boost the performance of the Intra-Human Relation Module. Even without bells and whistles, our method can compete or outperform current competition winners. We conduct extensive experiments on COCO, CrowdPose, and OCHuman datasets. The results demonstrate that the proposed model surpasses all the state-of-the-art methods. Concretely, the proposed method achieves 77.4% AP on CrowPose dataset and 67.8% AP on OCHuman dataset respectively, outperforming existing methods by a large margin. Additionally, the ablation study and visualization analysis also prove the effectiveness of our model.

Results

TaskDatasetMetricValueModel
Pose EstimationCOCO (Common Objects in Context)AP77.3I²R-Net (1st stage:HRFormer-B)
Pose EstimationCOCO (Common Objects in Context)AP5091I²R-Net (1st stage:HRFormer-B)
Pose EstimationCOCO (Common Objects in Context)AP7583.6I²R-Net (1st stage:HRFormer-B)
Pose EstimationCOCO (Common Objects in Context)APL84.5I²R-Net (1st stage:HRFormer-B)
Pose EstimationCOCO (Common Objects in Context)APM73I²R-Net (1st stage:HRFormer-B)
Pose EstimationCOCO (Common Objects in Context)AR82.1I²R-Net (1st stage:HRFormer-B)
Pose EstimationCrowdPoseAP Easy83.8I²R-Net (1st stage: HRFormer-B)
Pose EstimationCrowdPoseAP Hard69.3I²R-Net (1st stage: HRFormer-B)
Pose EstimationCrowdPoseAP Medium78.1I²R-Net (1st stage: HRFormer-B)
Pose EstimationCrowdPosemAP @0.5:0.9577.4I²R-Net (1st stage: HRFormer-B)
Pose EstimationOCHumanAP5085I²R-Net (1st stage:TransPose-H)
Pose EstimationOCHumanAP7572.8I²R-Net (1st stage:TransPose-H)
Pose EstimationOCHumanValidation AP67.8I²R-Net (1st stage:TransPose-H)
3DCOCO (Common Objects in Context)AP77.3I²R-Net (1st stage:HRFormer-B)
3DCOCO (Common Objects in Context)AP5091I²R-Net (1st stage:HRFormer-B)
3DCOCO (Common Objects in Context)AP7583.6I²R-Net (1st stage:HRFormer-B)
3DCOCO (Common Objects in Context)APL84.5I²R-Net (1st stage:HRFormer-B)
3DCOCO (Common Objects in Context)APM73I²R-Net (1st stage:HRFormer-B)
3DCOCO (Common Objects in Context)AR82.1I²R-Net (1st stage:HRFormer-B)
3DCrowdPoseAP Easy83.8I²R-Net (1st stage: HRFormer-B)
3DCrowdPoseAP Hard69.3I²R-Net (1st stage: HRFormer-B)
3DCrowdPoseAP Medium78.1I²R-Net (1st stage: HRFormer-B)
3DCrowdPosemAP @0.5:0.9577.4I²R-Net (1st stage: HRFormer-B)
3DOCHumanAP5085I²R-Net (1st stage:TransPose-H)
3DOCHumanAP7572.8I²R-Net (1st stage:TransPose-H)
3DOCHumanValidation AP67.8I²R-Net (1st stage:TransPose-H)
Multi-Person Pose EstimationCrowdPoseAP Easy83.8I²R-Net (1st stage: HRFormer-B)
Multi-Person Pose EstimationCrowdPoseAP Hard69.3I²R-Net (1st stage: HRFormer-B)
Multi-Person Pose EstimationCrowdPoseAP Medium78.1I²R-Net (1st stage: HRFormer-B)
Multi-Person Pose EstimationCrowdPosemAP @0.5:0.9577.4I²R-Net (1st stage: HRFormer-B)
Multi-Person Pose EstimationOCHumanAP5085I²R-Net (1st stage:TransPose-H)
Multi-Person Pose EstimationOCHumanAP7572.8I²R-Net (1st stage:TransPose-H)
Multi-Person Pose EstimationOCHumanValidation AP67.8I²R-Net (1st stage:TransPose-H)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)AP77.3I²R-Net (1st stage:HRFormer-B)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)AP5091I²R-Net (1st stage:HRFormer-B)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)AP7583.6I²R-Net (1st stage:HRFormer-B)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)APL84.5I²R-Net (1st stage:HRFormer-B)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)APM73I²R-Net (1st stage:HRFormer-B)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)AR82.1I²R-Net (1st stage:HRFormer-B)
1 Image, 2*2 StitchiCrowdPoseAP Easy83.8I²R-Net (1st stage: HRFormer-B)
1 Image, 2*2 StitchiCrowdPoseAP Hard69.3I²R-Net (1st stage: HRFormer-B)
1 Image, 2*2 StitchiCrowdPoseAP Medium78.1I²R-Net (1st stage: HRFormer-B)
1 Image, 2*2 StitchiCrowdPosemAP @0.5:0.9577.4I²R-Net (1st stage: HRFormer-B)
1 Image, 2*2 StitchiOCHumanAP5085I²R-Net (1st stage:TransPose-H)
1 Image, 2*2 StitchiOCHumanAP7572.8I²R-Net (1st stage:TransPose-H)
1 Image, 2*2 StitchiOCHumanValidation AP67.8I²R-Net (1st stage:TransPose-H)

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16