TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/POSTER++: A simpler and stronger facial expression recogni...

POSTER++: A simpler and stronger facial expression recognition network

Jiawei Mao, Rui Xu, Xuesong Yin, Yuanqi Chang, Binling Nie, Aibin Huang

2023-01-28Facial Expression RecognitionFacial Expression Recognition (FER)
PaperPDFCode(official)

Abstract

Facial expression recognition (FER) plays an important role in a variety of real-world applications such as human-computer interaction. POSTER achieves the state-of-the-art (SOTA) performance in FER by effectively combining facial landmark and image features through two-stream pyramid cross-fusion design. However, the architecture of POSTER is undoubtedly complex. It causes expensive computational costs. In order to relieve the computational pressure of POSTER, in this paper, we propose POSTER++. It improves POSTER in three directions: cross-fusion, two-stream, and multi-scale feature extraction. In cross-fusion, we use window-based cross-attention mechanism replacing vanilla cross-attention mechanism. We remove the image-to-landmark branch in the two-stream design. For multi-scale feature extraction, POSTER++ combines images with landmark's multi-scale features to replace POSTER's pyramid design. Extensive experiments on several standard datasets show that our POSTER++ achieves the SOTA FER performance with the minimum computational cost. For example, POSTER++ reached 92.21% on RAF-DB, 67.49% on AffectNet (7 cls) and 63.77% on AffectNet (8 cls), respectively, using only 8.4G floating point operations (FLOPs) and 43.7M parameters (Param). This demonstrates the effectiveness of our improvements.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingRAF-DBOverall Accuracy92.21POSTER++
Facial Recognition and ModellingAffectNetAccuracy (7 emotion)67.49POSTER++
Facial Recognition and ModellingAffectNetAccuracy (8 emotion)63.77POSTER++
Face ReconstructionRAF-DBOverall Accuracy92.21POSTER++
Face ReconstructionAffectNetAccuracy (7 emotion)67.49POSTER++
Face ReconstructionAffectNetAccuracy (8 emotion)63.77POSTER++
Facial Expression Recognition (FER)RAF-DBOverall Accuracy92.21POSTER++
Facial Expression Recognition (FER)AffectNetAccuracy (7 emotion)67.49POSTER++
Facial Expression Recognition (FER)AffectNetAccuracy (8 emotion)63.77POSTER++
3DRAF-DBOverall Accuracy92.21POSTER++
3DAffectNetAccuracy (7 emotion)67.49POSTER++
3DAffectNetAccuracy (8 emotion)63.77POSTER++
3D Face ModellingRAF-DBOverall Accuracy92.21POSTER++
3D Face ModellingAffectNetAccuracy (7 emotion)67.49POSTER++
3D Face ModellingAffectNetAccuracy (8 emotion)63.77POSTER++
3D Face ReconstructionRAF-DBOverall Accuracy92.21POSTER++
3D Face ReconstructionAffectNetAccuracy (7 emotion)67.49POSTER++
3D Face ReconstructionAffectNetAccuracy (8 emotion)63.77POSTER++

Related Papers

Multimodal Prompt Alignment for Facial Expression Recognition2025-06-26Enhancing Ambiguous Dynamic Facial Expression Recognition with Soft Label-based Data Augmentation2025-06-25Using Vision Language Models to Detect Students' Academic Emotion through Facial Expressions2025-06-12EfficientFER: EfficientNetv2 Based Deep Learning Approach for Facial Expression Recognition2025-06-02TKFNet: Learning Texture Key Factor Driven Feature for Facial Expression Recognition2025-05-15Unsupervised Multiview Contrastive Language-Image Joint Learning with Pseudo-Labeled Prompts Via Vision-Language Model for 3D/4D Facial Expression Recognition2025-05-14Achieving 3D Attention via Triplet Squeeze and Excitation Block2025-05-09Some Optimizers are More Equal: Understanding the Role of Optimizers in Group Fairness2025-04-21