TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Peak-Piloted Deep Network for Facial Expression Recognition

Peak-Piloted Deep Network for Facial Expression Recognition

Xiangyun Zhao, Xiaodan Liang, Luoqi Liu, Teng Li, Yugang Han, Nuno Vasconcelos, Shuicheng Yan

2016-07-24Face RecognitionRobust Face RecognitionFacial Expression RecognitionFacial Expression Recognition (FER)
PaperPDF

Abstract

Objective functions for training of deep networks for face-related recognition tasks, such as facial expression recognition (FER), usually consider each sample independently. In this work, we present a novel peak-piloted deep network (PPDN) that uses a sample with peak expression (easy sample) to supervise the intermediate feature responses for a sample of non-peak expression (hard sample) of the same type and from the same subject. The expression evolving process from non-peak expression to peak expression can thus be implicitly embedded in the network to achieve the invariance to expression intensities. A special purpose back-propagation procedure, peak gradient suppression (PGS), is proposed for network training. It drives the intermediate-layer feature responses of non-peak expression samples towards those of the corresponding peak expression samples, while avoiding the inverse. This avoids degrading the recognition capability for samples of peak expression due to interference from their non-peak expression counterparts. Extensive comparisons on two popular FER datasets, Oulu-CASIA and CK+, demonstrate the superiority of the PPDN over state-ofthe-art FER methods, as well as the advantages of both the network structure and the optimization strategy. Moreover, it is shown that PPDN is a general architecture, extensible to other tasks by proper definition of peak and non-peak samples. This is validated by experiments that show state-of-the-art performance on pose-invariant face recognition, using the Multi-PIE dataset.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingOulu-CASIAAccuracy (10-fold)84.59PPDN
Face ReconstructionOulu-CASIAAccuracy (10-fold)84.59PPDN
Facial Expression Recognition (FER)Oulu-CASIAAccuracy (10-fold)84.59PPDN
3DOulu-CASIAAccuracy (10-fold)84.59PPDN
3D Face ModellingOulu-CASIAAccuracy (10-fold)84.59PPDN
3D Face ReconstructionOulu-CASIAAccuracy (10-fold)84.59PPDN

Related Papers

ProxyFusion: Face Feature Aggregation Through Sparse Experts2025-09-24Non-Adaptive Adversarial Face Generation2025-07-16Attributes Shape the Embedding Space of Face Recognition Models2025-07-15Face mask detection project report.2025-07-02Multimodal Prompt Alignment for Facial Expression Recognition2025-06-26On the Burstiness of Faces in Set2025-06-25Enhancing Ambiguous Dynamic Facial Expression Recognition with Soft Label-based Data Augmentation2025-06-25Identifying Physically Realizable Triggers for Backdoored Face Recognition Networks2025-06-24