TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Zolly: Zoom Focal Length Correctly for Perspective-Distort...

Zolly: Zoom Focal Length Correctly for Perspective-Distorted Human Mesh Reconstruction

Wenjia Wang, Yongtao Ge, Haiyi Mei, Zhongang Cai, Qingping Sun, Yanjun Wang, Chunhua Shen, Lei Yang, Taku Komura

2023-03-24ICCV 2023 13D Human Pose Estimation3D Reconstruction
PaperPDFCode(official)

Abstract

As it is hard to calibrate single-view RGB images in the wild, existing 3D human mesh reconstruction (3DHMR) methods either use a constant large focal length or estimate one based on the background environment context, which can not tackle the problem of the torso, limb, hand or face distortion caused by perspective camera projection when the camera is close to the human body. The naive focal length assumptions can harm this task with the incorrectly formulated projection matrices. To solve this, we propose Zolly, the first 3DHMR method focusing on perspective-distorted images. Our approach begins with analysing the reason for perspective distortion, which we find is mainly caused by the relative location of the human body to the camera center. We propose a new camera model and a novel 2D representation, termed distortion image, which describes the 2D dense distortion scale of the human body. We then estimate the distance from distortion scale features rather than environment context features. Afterwards, we integrate the distortion feature with image features to reconstruct the body mesh. To formulate the correct projection matrix and locate the human body position, we simultaneously use perspective and weak-perspective projection loss. Since existing datasets could not handle this task, we propose the first synthetic dataset PDHuman and extend two real-world datasets tailored for this task, all containing perspective-distorted human images. Extensive experiments show that Zolly outperforms existing state-of-the-art methods on both perspective-distorted datasets and the standard benchmark (3DPW).

Results

TaskDatasetMetricValueModel
3D Human Pose Estimation3DPWMPJPE65Zolly (HRNet-w48)
3D Human Pose Estimation3DPWMPVPE76.3Zolly (HRNet-w48)
3D Human Pose Estimation3DPWPA-MPJPE39.8Zolly (HRNet-w48)
Pose Estimation3DPWMPJPE65Zolly (HRNet-w48)
Pose Estimation3DPWMPVPE76.3Zolly (HRNet-w48)
Pose Estimation3DPWPA-MPJPE39.8Zolly (HRNet-w48)
3D3DPWMPJPE65Zolly (HRNet-w48)
3D3DPWMPVPE76.3Zolly (HRNet-w48)
3D3DPWPA-MPJPE39.8Zolly (HRNet-w48)
1 Image, 2*2 Stitchi3DPWMPJPE65Zolly (HRNet-w48)
1 Image, 2*2 Stitchi3DPWMPVPE76.3Zolly (HRNet-w48)
1 Image, 2*2 Stitchi3DPWPA-MPJPE39.8Zolly (HRNet-w48)

Related Papers

AutoPartGen: Autogressive 3D Part Generation and Discovery2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16BRUM: Robust 3D Vehicle Reconstruction from 360 Sparse Images2025-07-16Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Binomial Self-Compensation: Mechanism and Suppression of Motion Error in Phase-Shifting Profilometry2025-07-14An Efficient Approach for Muscle Segmentation and 3D Reconstruction Using Keypoint Tracking in MRI Scan2025-07-11Review of Feed-forward 3D Reconstruction: From DUSt3R to VGGT2025-07-11DreamGrasp: Zero-Shot 3D Multi-Object Reconstruction from Partial-View Images for Robotic Manipulation2025-07-08