TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/HTNet for micro-expression recognition

HTNet for micro-expression recognition

Zhifeng Wang, Kaihao Zhang, Wenhan Luo, Ramesh Sankaranarayana

2023-07-27Facial Emotion RecognitionMicro Expression RecognitionMicro-Expression RecognitionEmotion Recognition
PaperPDFCode(official)

Abstract

Facial expression is related to facial muscle contractions and different muscle movements correspond to different emotional states. For micro-expression recognition, the muscle movements are usually subtle, which has a negative impact on the performance of current facial emotion recognition algorithms. Most existing methods use self-attention mechanisms to capture relationships between tokens in a sequence, but they do not take into account the inherent spatial relationships between facial landmarks. This can result in sub-optimal performance on micro-expression recognition tasks.Therefore, learning to recognize facial muscle movements is a key challenge in the area of micro-expression recognition. In this paper, we propose a Hierarchical Transformer Network (HTNet) to identify critical areas of facial muscle movement. HTNet includes two major components: a transformer layer that leverages the local temporal features and an aggregation layer that extracts local and global semantical facial features. Specifically, HTNet divides the face into four different facial areas: left lip area, left eye area, right eye area and right lip area. The transformer layer is used to focus on representing local minor muscle movement with local self-attention in each area. The aggregation layer is used to learn the interactions between eye areas and lip areas. The experiments on four publicly available micro-expression datasets show that the proposed approach outperforms previous methods by a large margin. The codes and models are available at: \url{https://github.com/wangzhifengharrison/HTNet}

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingCASME IIUAR95.16HTNet
Facial Recognition and ModellingCASME IIUF195.32HTNet
Face ReconstructionCASME IIUAR95.16HTNet
Face ReconstructionCASME IIUF195.32HTNet
Facial Expression Recognition (FER)CASME IIUAR95.16HTNet
Facial Expression Recognition (FER)CASME IIUF195.32HTNet
3DCASME IIUAR95.16HTNet
3DCASME IIUF195.32HTNet
3D Face ModellingCASME IIUAR95.16HTNet
3D Face ModellingCASME IIUF195.32HTNet
3D Face ReconstructionCASME IIUAR95.16HTNet
3D Face ReconstructionCASME IIUF195.32HTNet

Related Papers

Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation2025-07-21Camera-based implicit mind reading by capturing higher-order semantic dynamics of human gaze within environmental context2025-07-17A Robust Incomplete Multimodal Low-Rank Adaptation Approach for Emotion Recognition2025-07-15Dynamic Parameter Memory: Temporary LoRA-Enhanced LLM for Long-Sequence Emotion Recognition in Conversation2025-07-11CAST-Phys: Contactless Affective States Through Physiological signals Database2025-07-08Exploring Remote Physiological Signal Measurement under Dynamic Lighting Conditions at Night: Dataset, Experiment, and Analysis2025-07-06How to Retrieve Examples in In-context Learning to Improve Conversational Emotion Recognition using Large Language Models?2025-06-25MATER: Multi-level Acoustic and Textual Emotion Representation for Interpretable Speech Emotion Recognition2025-06-24