TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Distract Your Attention: Multi-head Cross Attention Networ...

Distract Your Attention: Multi-head Cross Attention Network for Facial Expression Recognition

Zhengyao Wen, Wenzhong Lin, Tao Wang, Ge Xu

2021-09-15Facial Expression RecognitionFacial Expression Recognition (FER)
PaperPDFCodeCode(official)

Abstract

We present a novel facial expression recognition network, called Distract your Attention Network (DAN). Our method is based on two key observations. Firstly, multiple classes share inherently similar underlying facial appearance, and their differences could be subtle. Secondly, facial expressions exhibit themselves through multiple facial regions simultaneously, and the recognition requires a holistic approach by encoding high-order interactions among local features. To address these issues, we propose our DAN with three key components: Feature Clustering Network (FCN), Multi-head cross Attention Network (MAN), and Attention Fusion Network (AFN). The FCN extracts robust features by adopting a large-margin learning objective to maximize class separability. In addition, the MAN instantiates a number of attention heads to simultaneously attend to multiple facial areas and build attention maps on these regions. Further, the AFN distracts these attentions to multiple locations before fusing the attention maps to a comprehensive one. Extensive experiments on three public datasets (including AffectNet, RAF-DB, and SFEW 2.0) verified that the proposed method consistently achieves state-of-the-art facial expression recognition performance. Code will be made available at https://github.com/yaoing/DAN.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingRAF-DBOverall Accuracy89.7DAN
Facial Recognition and ModellingAffectNetAccuracy (7 emotion)65.69DAN
Facial Recognition and ModellingAffectNetAccuracy (8 emotion)62.09DAN
Face ReconstructionRAF-DBOverall Accuracy89.7DAN
Face ReconstructionAffectNetAccuracy (7 emotion)65.69DAN
Face ReconstructionAffectNetAccuracy (8 emotion)62.09DAN
Facial Expression Recognition (FER)RAF-DBOverall Accuracy89.7DAN
Facial Expression Recognition (FER)AffectNetAccuracy (7 emotion)65.69DAN
Facial Expression Recognition (FER)AffectNetAccuracy (8 emotion)62.09DAN
3DRAF-DBOverall Accuracy89.7DAN
3DAffectNetAccuracy (7 emotion)65.69DAN
3DAffectNetAccuracy (8 emotion)62.09DAN
3D Face ModellingRAF-DBOverall Accuracy89.7DAN
3D Face ModellingAffectNetAccuracy (7 emotion)65.69DAN
3D Face ModellingAffectNetAccuracy (8 emotion)62.09DAN
3D Face ReconstructionRAF-DBOverall Accuracy89.7DAN
3D Face ReconstructionAffectNetAccuracy (7 emotion)65.69DAN
3D Face ReconstructionAffectNetAccuracy (8 emotion)62.09DAN

Related Papers

Multimodal Prompt Alignment for Facial Expression Recognition2025-06-26Enhancing Ambiguous Dynamic Facial Expression Recognition with Soft Label-based Data Augmentation2025-06-25Using Vision Language Models to Detect Students' Academic Emotion through Facial Expressions2025-06-12EfficientFER: EfficientNetv2 Based Deep Learning Approach for Facial Expression Recognition2025-06-02TKFNet: Learning Texture Key Factor Driven Feature for Facial Expression Recognition2025-05-15Unsupervised Multiview Contrastive Language-Image Joint Learning with Pseudo-Labeled Prompts Via Vision-Language Model for 3D/4D Facial Expression Recognition2025-05-14Achieving 3D Attention via Triplet Squeeze and Excitation Block2025-05-09Some Optimizers are More Equal: Understanding the Role of Optimizers in Group Fairness2025-04-21