TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/DSFNet: Dual Space Fusion Network for Occlusion-Robust 3D ...

DSFNet: Dual Space Fusion Network for Occlusion-Robust 3D Dense Face Alignment

Heyuan Li, Bo wang, Yu Cheng, Mohan Kankanhalli, Robby T. Tan

2023-05-19CVPR 2023 1Face Alignment3D Face ReconstructionHead Pose Estimation
PaperPDFCode(official)

Abstract

Sensitivity to severe occlusion and large view angles limits the usage scenarios of the existing monocular 3D dense face alignment methods. The state-of-the-art 3DMM-based method, directly regresses the model's coefficients, underutilizing the low-level 2D spatial and semantic information, which can actually offer cues for face shape and orientation. In this work, we demonstrate how modeling 3D facial geometry in image and model space jointly can solve the occlusion and view angle problems. Instead of predicting the whole face directly, we regress image space features in the visible facial region by dense prediction first. Subsequently, we predict our model's coefficients based on the regressed feature of the visible regions, leveraging the prior knowledge of whole face geometry from the morphable models to complete the invisible regions. We further propose a fusion network that combines the advantages of both the image and model space predictions to achieve high robustness and accuracy in unconstrained scenarios. Thanks to the proposed fusion module, our method is robust not only to occlusion and large pitch and roll view angles, which is the benefit of our image space approach, but also to noise and large yaw angles, which is the benefit of our model space method. Comprehensive evaluations demonstrate the superior performance of our method compared with the state-of-the-art methods. On the 3D dense face alignment task, we achieve 3.80% NME on the AFLW2000-3D dataset, which outperforms the state-of-the-art method by 5.5%. Code is available at https://github.com/lhyfst/DSFNet.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingAFLW2000-3DMean NME3.16DSFNet-is
Pose EstimationAFLW2000MAE3.25DSFNet-f
Face ReconstructionAFLW2000-3DMean NME3.16DSFNet-is
3DAFLW2000MAE3.25DSFNet-f
3DAFLW2000-3DMean NME3.16DSFNet-is
3D Face ModellingAFLW2000-3DMean NME3.16DSFNet-is
3D Face ReconstructionAFLW2000-3DMean NME3.16DSFNet-is
1 Image, 2*2 StitchiAFLW2000MAE3.25DSFNet-f

Related Papers

From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17Towards Large-Scale Pose-Invariant Face Recognition Using Face Defrontalization2025-06-04Event-based Egocentric Human Pose Estimation in Dynamic Environment2025-05-28HonestFace: Towards Honest Face Restoration with One-Step Diffusion Model2025-05-243D Face Reconstruction Error Decomposed: A Modular Benchmark for Fair and Fast Method Evaluation2025-05-23Multimodal Emotion Coupling via Speech-to-Facial and Bodily Gestures in Dyadic Interaction2025-05-08Pixel3DMM: Versatile Screen-Space Priors for Single-Image 3D Face Reconstruction2025-05-01DMAGaze: Gaze Estimation Based on Feature Disentanglement and Multi-Scale Attention2025-04-15