TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Fine-Grained Head Pose Estimation Without Keypoints

Fine-Grained Head Pose Estimation Without Keypoints

Nataniel Ruiz, Eunji Chong, James M. Rehg

2017-10-02Face AlignmentPose EstimationGaze EstimationHead Pose Estimation
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCode

Abstract

Estimating the head pose of a person is a crucial problem that has a large amount of applications such as aiding in gaze estimation, modeling attention, fitting 3D models to video and performing face alignment. Traditionally head pose is computed by estimating some keypoints from the target face and solving the 2D to 3D correspondence problem with a mean human head model. We argue that this is a fragile method because it relies entirely on landmark detection performance, the extraneous head model and an ad-hoc fitting step. We present an elegant and robust way to determine pose by training a multi-loss convolutional neural network on 300W-LP, a large synthetically expanded dataset, to predict intrinsic Euler angles (yaw, pitch and roll) directly from image intensities through joint binned pose classification and regression. We present empirical tests on common in-the-wild pose benchmark datasets which show state-of-the-art results. Additionally we test our method on a dataset usually used for pose estimation using depth and start to close the gap with state-of-the-art depth pose methods. We open-source our training and testing code as well as release our pre-trained models.

Results

TaskDatasetMetricValueModel
Pose EstimationAFLW2000Geodesic Error (GE)993Hopenet
Pose EstimationAFLW2000MAE6.15Hopenet
Pose EstimationAFLW2000MAE6.155Multi-Loss ResNet50 (a=2)
Pose EstimationBIWIGeodesic Error (GE)9.53hopenet
Pose EstimationBIWIGeodesic Error - aligned (GE)6.6hopenet
Pose EstimationBIWIMAE (trained with other data)4.89hopenet
Pose EstimationBIWIMAE-aligned (trained with other data)3.48hopenet
Pose EstimationBIWIMAE (trained with BIWI data)4.895Multi-Loss ResNet50
Pose EstimationAFLWMAE5.324Ruiz et al.
3DAFLW2000Geodesic Error (GE)993Hopenet
3DAFLW2000MAE6.15Hopenet
3DAFLW2000MAE6.155Multi-Loss ResNet50 (a=2)
3DBIWIGeodesic Error (GE)9.53hopenet
3DBIWIGeodesic Error - aligned (GE)6.6hopenet
3DBIWIMAE (trained with other data)4.89hopenet
3DBIWIMAE-aligned (trained with other data)3.48hopenet
3DBIWIMAE (trained with BIWI data)4.895Multi-Loss ResNet50
3DAFLWMAE5.324Ruiz et al.
1 Image, 2*2 StitchiAFLW2000Geodesic Error (GE)993Hopenet
1 Image, 2*2 StitchiAFLW2000MAE6.15Hopenet
1 Image, 2*2 StitchiAFLW2000MAE6.155Multi-Loss ResNet50 (a=2)
1 Image, 2*2 StitchiBIWIGeodesic Error (GE)9.53hopenet
1 Image, 2*2 StitchiBIWIGeodesic Error - aligned (GE)6.6hopenet
1 Image, 2*2 StitchiBIWIMAE (trained with other data)4.89hopenet
1 Image, 2*2 StitchiBIWIMAE-aligned (trained with other data)3.48hopenet
1 Image, 2*2 StitchiBIWIMAE (trained with BIWI data)4.895Multi-Loss ResNet50
1 Image, 2*2 StitchiAFLWMAE5.324Ruiz et al.

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16