TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Region Attention Networks for Pose and Occlusion Robust Fa...

Region Attention Networks for Pose and Occlusion Robust Facial Expression Recognition

Kai Wang, Xiaojiang Peng, Jianfei Yang, Debin Meng, Yu Qiao

2019-05-10Facial Expression RecognitionFacial Expression Recognition (FER)
PaperPDFCode(official)

Abstract

Occlusion and pose variations, which can change facial appearance significantly, are two major obstacles for automatic Facial Expression Recognition (FER). Though automatic FER has made substantial progresses in the past few decades, occlusion-robust and pose-invariant issues of FER have received relatively less attention, especially in real-world scenarios. This paper addresses the real-world pose and occlusion robust FER problem with three-fold contributions. First, to stimulate the research of FER under real-world occlusions and variant poses, we build several in-the-wild facial expression datasets with manual annotations for the community. Second, we propose a novel Region Attention Network (RAN), to adaptively capture the importance of facial regions for occlusion and pose variant FER. The RAN aggregates and embeds varied number of region features produced by a backbone convolutional neural network into a compact fixed-length representation. Last, inspired by the fact that facial expressions are mainly defined by facial action units, we propose a region biased loss to encourage high attention weights for the most important regions. We validate our RAN and region biased loss on both our built test datasets and four popular datasets: FERPlus, AffectNet, RAF-DB, and SFEW. Extensive experiments show that our RAN and region biased loss largely improve the performance of FER with occlusion and variant pose. Our method also achieves state-of-the-art results on FERPlus, AffectNet, RAF-DB, and SFEW. Code and the collected test data will be publicly available.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingRAF-DBOverall Accuracy86.9RAN (ResNet-18)
Facial Recognition and ModellingSFEWAccuracy56.4RAN (VGG16+ResNet18)
Facial Recognition and ModellingAffectNetAccuracy (8 emotion)59.5RAN (ResNet-18+)
Facial Recognition and ModellingFERPlusAccuracy(pretrained)89.16RAN (VGG-16)
Face ReconstructionRAF-DBOverall Accuracy86.9RAN (ResNet-18)
Face ReconstructionSFEWAccuracy56.4RAN (VGG16+ResNet18)
Face ReconstructionAffectNetAccuracy (8 emotion)59.5RAN (ResNet-18+)
Face ReconstructionFERPlusAccuracy(pretrained)89.16RAN (VGG-16)
Facial Expression Recognition (FER)RAF-DBOverall Accuracy86.9RAN (ResNet-18)
Facial Expression Recognition (FER)SFEWAccuracy56.4RAN (VGG16+ResNet18)
Facial Expression Recognition (FER)FERPlusAccuracy(pretrained)89.16RAN (VGG-16)
Facial Expression Recognition (FER)AffectNetAccuracy (8 emotion)59.5RAN (ResNet-18+)
3DRAF-DBOverall Accuracy86.9RAN (ResNet-18)
3DSFEWAccuracy56.4RAN (VGG16+ResNet18)
3DAffectNetAccuracy (8 emotion)59.5RAN (ResNet-18+)
3DFERPlusAccuracy(pretrained)89.16RAN (VGG-16)
3D Face ModellingRAF-DBOverall Accuracy86.9RAN (ResNet-18)
3D Face ModellingSFEWAccuracy56.4RAN (VGG16+ResNet18)
3D Face ModellingFERPlusAccuracy(pretrained)89.16RAN (VGG-16)
3D Face ModellingAffectNetAccuracy (8 emotion)59.5RAN (ResNet-18+)
3D Face ReconstructionRAF-DBOverall Accuracy86.9RAN (ResNet-18)
3D Face ReconstructionSFEWAccuracy56.4RAN (VGG16+ResNet18)
3D Face ReconstructionAffectNetAccuracy (8 emotion)59.5RAN (ResNet-18+)
3D Face ReconstructionFERPlusAccuracy(pretrained)89.16RAN (VGG-16)

Related Papers

Multimodal Prompt Alignment for Facial Expression Recognition2025-06-26Enhancing Ambiguous Dynamic Facial Expression Recognition with Soft Label-based Data Augmentation2025-06-25Using Vision Language Models to Detect Students' Academic Emotion through Facial Expressions2025-06-12EfficientFER: EfficientNetv2 Based Deep Learning Approach for Facial Expression Recognition2025-06-02TKFNet: Learning Texture Key Factor Driven Feature for Facial Expression Recognition2025-05-15Unsupervised Multiview Contrastive Language-Image Joint Learning with Pseudo-Labeled Prompts Via Vision-Language Model for 3D/4D Facial Expression Recognition2025-05-14Achieving 3D Attention via Triplet Squeeze and Excitation Block2025-05-09Some Optimizers are More Equal: Understanding the Role of Optimizers in Group Fairness2025-04-21