TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Expression, Affect, Action Unit Recognition: Aff-Wild2, Mu...

Expression, Affect, Action Unit Recognition: Aff-Wild2, Multi-Task Learning and ArcFace

Dimitrios Kollias, Stefanos Zafeiriou

2019-09-25Multi-Task LearningFacial Expression Recognition (FER)Arousal EstimationEmotion RecognitionAction Unit Detection
PaperPDF

Abstract

Affective computing has been largely limited in terms of available data resources. The need to collect and annotate diverse in-the-wild datasets has become apparent with the rise of deep learning models, as the default approach to address any computer vision task. Some in-the-wild databases have been recently proposed. However: i) their size is small, ii) they are not audiovisual, iii) only a small part is manually annotated, iv) they contain a small number of subjects, or v) they are not annotated for all main behavior tasks (valence-arousal estimation, action unit detection and basic expression classification). To address these, we substantially extend the largest available in-the-wild database (Aff-Wild) to study continuous emotions such as valence and arousal. Furthermore, we annotate parts of the database with basic expressions and action units. As a consequence, for the first time, this allows the joint study of all three types of behavior states. We call this database Aff-Wild2. We conduct extensive experiments with CNN and CNN-RNN architectures that use visual and audio modalities; these networks are trained on Aff-Wild2 and their performance is then evaluated on 10 publicly available emotion databases. We show that the networks achieve state-of-the-art performance for the emotion recognition tasks. Additionally, we adapt the ArcFace loss function in the emotion recognition context and use it for training two new networks on Aff-Wild2 and then re-train them in a variety of diverse expression recognition databases. The networks are shown to improve the existing state-of-the-art. The database, emotion recognition models and source code are available at http://ibug.doc.ic.ac.uk/resources/aff-wild2.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingRAF-DBAvg. Accuracy76MT-ArcVGG
Facial Recognition and ModellingAffectNetAccuracy (8 emotion)63MT-ArcRes
Face ReconstructionRAF-DBAvg. Accuracy76MT-ArcVGG
Face ReconstructionAffectNetAccuracy (8 emotion)63MT-ArcRes
Facial Expression Recognition (FER)RAF-DBAvg. Accuracy76MT-ArcVGG
Facial Expression Recognition (FER)AffectNetAccuracy (8 emotion)63MT-ArcRes
3DRAF-DBAvg. Accuracy76MT-ArcVGG
3DAffectNetAccuracy (8 emotion)63MT-ArcRes
3D Face ModellingRAF-DBAvg. Accuracy76MT-ArcVGG
3D Face ModellingAffectNetAccuracy (8 emotion)63MT-ArcRes
3D Face ReconstructionRAF-DBAvg. Accuracy76MT-ArcVGG
3D Face ReconstructionAffectNetAccuracy (8 emotion)63MT-ArcRes

Related Papers

Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation2025-07-21SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Camera-based implicit mind reading by capturing higher-order semantic dynamics of human gaze within environmental context2025-07-17Robust-Multi-Task Gradient Boosting2025-07-15A Robust Incomplete Multimodal Low-Rank Adaptation Approach for Emotion Recognition2025-07-15Dynamic Parameter Memory: Temporary LoRA-Enhanced LLM for Long-Sequence Emotion Recognition in Conversation2025-07-11SAMO: A Lightweight Sharpness-Aware Approach for Multi-Task Optimization with Joint Global-Local Perturbation2025-07-10CAST-Phys: Contactless Affective States Through Physiological signals Database2025-07-08