TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/EMOCA: Emotion Driven Monocular Face Capture and Animation

EMOCA: Emotion Driven Monocular Face Capture and Animation

Radek Danecek, Michael J. Black, Timo Bolkart

2022-04-24CVPR 2022 13D geometryFace RecognitionFace Reconstruction3D Reconstruction3D Face ReconstructionEmotion Recognition
PaperPDFCode(official)

Abstract

As 3D facial avatars become more widely used for communication, it is critical that they faithfully convey emotion. Unfortunately, the best recent methods that regress parametric 3D face models from monocular images are unable to capture the full spectrum of facial expression, such as subtle or extreme emotions. We find the standard reconstruction metrics used for training (landmark reprojection error, photometric error, and face recognition loss) are insufficient to capture high-fidelity expressions. The result is facial geometries that do not match the emotional content of the input image. We address this with EMOCA (EMOtion Capture and Animation), by introducing a novel deep perceptual emotion consistency loss during training, which helps ensure that the reconstructed 3D expression matches the expression depicted in the input image. While EMOCA achieves 3D reconstruction errors that are on par with the current best methods, it significantly outperforms them in terms of the quality of the reconstructed expression and the perceived emotional content. We also directly regress levels of valence and arousal and classify basic expressions from the estimated 3D face parameters. On the task of in-the-wild emotion recognition, our purely geometric approach is on par with the best image-based methods, highlighting the value of 3D geometry in analyzing human behavior. The model and code are publicly available at https://emoca.is.tue.mpg.de.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingREALYall2.103EMOCA-c
Facial Recognition and ModellingREALYall2.388EMOCA-f
Facial Recognition and ModellingREALY (side-view)all2.125EMOCA-c
Facial Recognition and ModellingREALY (side-view)all2.402EMOCA-f
Face ReconstructionREALYall2.103EMOCA-c
Face ReconstructionREALYall2.388EMOCA-f
Face ReconstructionREALY (side-view)all2.125EMOCA-c
Face ReconstructionREALY (side-view)all2.402EMOCA-f
3DREALYall2.103EMOCA-c
3DREALYall2.388EMOCA-f
3DREALY (side-view)all2.125EMOCA-c
3DREALY (side-view)all2.402EMOCA-f
3D Face ModellingREALYall2.103EMOCA-c
3D Face ModellingREALYall2.388EMOCA-f
3D Face ModellingREALY (side-view)all2.125EMOCA-c
3D Face ModellingREALY (side-view)all2.402EMOCA-f
3D Face ReconstructionREALYall2.103EMOCA-c
3D Face ReconstructionREALYall2.388EMOCA-f
3D Face ReconstructionREALY (side-view)all2.125EMOCA-c
3D Face ReconstructionREALY (side-view)all2.402EMOCA-f

Related Papers

ProxyFusion: Face Feature Aggregation Through Sparse Experts2025-09-24Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation2025-07-21AutoPartGen: Autogressive 3D Part Generation and Discovery2025-07-17Camera-based implicit mind reading by capturing higher-order semantic dynamics of human gaze within environmental context2025-07-17Non-Adaptive Adversarial Face Generation2025-07-16SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16BRUM: Robust 3D Vehicle Reconstruction from 360 Sparse Images2025-07-16Robust 3D-Masked Part-level Editing in 3D Gaussian Splatting with Regularized Score Distillation Sampling2025-07-15