TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Style Aggregated Network for Facial Landmark Detection

Style Aggregated Network for Facial Landmark Detection

Xuanyi Dong, Yan Yan, Wanli Ouyang, Yi Yang

2018-03-12CVPR 2018 6Face AlignmentFacial Landmark Detection
PaperPDFCode(official)

Abstract

Recent advances in facial landmark detection achieve success by learning discriminative features from rich deformation of face shapes and poses. Besides the variance of faces themselves, the intrinsic variance of image styles, e.g., grayscale vs. color images, light vs. dark, intense vs. dull, and so on, has constantly been overlooked. This issue becomes inevitable as increasing web images are collected from various sources for training neural networks. In this work, we propose a style-aggregated approach to deal with the large intrinsic variance of image styles for facial landmark detection. Our method transforms original face images to style-aggregated images by a generative adversarial module. The proposed scheme uses the style-aggregated image to maintain face images that are more robust to environmental changes. Then the original face images accompanying with style-aggregated ones play a duet to train a landmark detector which is complementary to each other. In this way, for each face, our method takes two images as input, i.e., one in its original style and the other in the aggregated style. In experiments, we observe that the large variance of image styles would degenerate the performance of facial landmark detectors. Moreover, we show the robustness of our method to the large variance of image styles by comparing to a variant of our approach, in which the generative adversarial module is removed, and no style-aggregated images are used. Our approach is demonstrated to perform well when compared with state-of-the-art algorithms on benchmark datasets AFLW and 300-W. Code is publicly available on GitHub: https://github.com/D-X-Y/SAN

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingAFLW-19AUC_box@0.07 (%, Full)54SAN
Facial Recognition and ModellingAFLW-19NME_box (%, Full)4.04SAN
Facial Recognition and ModellingAFLW-19NME_diag (%, Frontal)1.85SAN
Facial Recognition and ModellingAFLW-19NME_diag (%, Full)1.91SAN
Facial Recognition and Modelling300WNME_inter-ocular (%, Challenge)6.6SAN
Facial Recognition and Modelling300WNME_inter-ocular (%, Common)3.34SAN
Facial Recognition and Modelling300WNME_inter-ocular (%, Full)3.98SAN
Facial Recognition and ModellingAFLW-FrontMean NME 1.85SAN
Facial Recognition and Modelling300WNME3.98SAN GT
Facial Recognition and ModellingAFLW-FullMean NME 1.91SAN
Facial Landmark DetectionAFLW-FrontMean NME 1.85SAN
Facial Landmark Detection300WNME3.98SAN GT
Facial Landmark DetectionAFLW-FullMean NME 1.91SAN
Face Reconstruction300WNME_inter-ocular (%, Challenge)6.6SAN
Face Reconstruction300WNME_inter-ocular (%, Common)3.34SAN
Face Reconstruction300WNME_inter-ocular (%, Full)3.98SAN
Face ReconstructionAFLW-19AUC_box@0.07 (%, Full)54SAN
Face ReconstructionAFLW-19NME_box (%, Full)4.04SAN
Face ReconstructionAFLW-19NME_diag (%, Frontal)1.85SAN
Face ReconstructionAFLW-19NME_diag (%, Full)1.91SAN
Face ReconstructionAFLW-FrontMean NME 1.85SAN
Face Reconstruction300WNME3.98SAN GT
Face ReconstructionAFLW-FullMean NME 1.91SAN
3D300WNME_inter-ocular (%, Challenge)6.6SAN
3D300WNME_inter-ocular (%, Common)3.34SAN
3D300WNME_inter-ocular (%, Full)3.98SAN
3DAFLW-19AUC_box@0.07 (%, Full)54SAN
3DAFLW-19NME_box (%, Full)4.04SAN
3DAFLW-19NME_diag (%, Frontal)1.85SAN
3DAFLW-19NME_diag (%, Full)1.91SAN
3DAFLW-FrontMean NME 1.85SAN
3D300WNME3.98SAN GT
3DAFLW-FullMean NME 1.91SAN
3D Face ModellingAFLW-19AUC_box@0.07 (%, Full)54SAN
3D Face ModellingAFLW-19NME_box (%, Full)4.04SAN
3D Face ModellingAFLW-19NME_diag (%, Frontal)1.85SAN
3D Face ModellingAFLW-19NME_diag (%, Full)1.91SAN
3D Face Modelling300WNME_inter-ocular (%, Challenge)6.6SAN
3D Face Modelling300WNME_inter-ocular (%, Common)3.34SAN
3D Face Modelling300WNME_inter-ocular (%, Full)3.98SAN
3D Face ModellingAFLW-FrontMean NME 1.85SAN
3D Face Modelling300WNME3.98SAN GT
3D Face ModellingAFLW-FullMean NME 1.91SAN
3D Face ReconstructionAFLW-19AUC_box@0.07 (%, Full)54SAN
3D Face ReconstructionAFLW-19NME_box (%, Full)4.04SAN
3D Face ReconstructionAFLW-19NME_diag (%, Frontal)1.85SAN
3D Face ReconstructionAFLW-19NME_diag (%, Full)1.91SAN
3D Face Reconstruction300WNME_inter-ocular (%, Challenge)6.6SAN
3D Face Reconstruction300WNME_inter-ocular (%, Common)3.34SAN
3D Face Reconstruction300WNME_inter-ocular (%, Full)3.98SAN
3D Face ReconstructionAFLW-FrontMean NME 1.85SAN
3D Face Reconstruction300WNME3.98SAN GT
3D Face ReconstructionAFLW-FullMean NME 1.91SAN

Related Papers

MOL: Joint Estimation of Micro-Expression, Optical Flow, and Landmark via Transformer-Graph-Style Convolution2025-06-17Towards Large-Scale Pose-Invariant Face Recognition Using Face Defrontalization2025-06-04HonestFace: Towards Honest Face Restoration with One-Step Diffusion Model2025-05-24Multimodal Emotion Coupling via Speech-to-Facial and Bodily Gestures in Dyadic Interaction2025-05-08Semantic Style Transfer for Enhancing Animal Facial Landmark Detection2025-05-08SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users2025-04-14Mitigating Knowledge Discrepancies among Multiple Datasets for Task-agnostic Unified Face Alignment2025-03-28Learning Person-Specific Animatable Face Models from In-the-Wild Images via a Shared Base Model2025-01-01