TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Wasserstein CNN: Learning Invariant Features for NIR-VIS F...

Wasserstein CNN: Learning Invariant Features for NIR-VIS Face Recognition

Ran He, Xiang Wu, Zhenan Sun, Tieniu Tan

2017-08-08Face RecognitionHeterogeneous Face Recognition
PaperPDF

Abstract

Heterogeneous face recognition (HFR) aims to match facial images acquired from different sensing modalities with mission-critical applications in forensics, security and commercial sectors. However, HFR is a much more challenging problem than traditional face recognition because of large intra-class variations of heterogeneous face images and limited training samples of cross-modality face image pairs. This paper proposes a novel approach namely Wasserstein CNN (convolutional neural networks, or WCNN for short) to learn invariant features between near-infrared and visual face images (i.e. NIR-VIS face recognition). The low-level layers of WCNN are trained with widely available face images in visual spectrum. The high-level layer is divided into three parts, i.e., NIR layer, VIS layer and NIR-VIS shared layer. The first two layers aims to learn modality-specific features and NIR-VIS shared layer is designed to learn modality-invariant feature subspace. Wasserstein distance is introduced into NIR-VIS shared layer to measure the dissimilarity between heterogeneous feature distributions. So W-CNN learning aims to achieve the minimization of Wasserstein distance between NIR distribution and VIS distribution for invariant deep feature representation of heterogeneous face images. To avoid the over-fitting problem on small-scale heterogeneous face data, a correlation prior is introduced on the fully-connected layers of WCNN network to reduce parameter space. This prior is implemented by a low-rank constraint in an end-to-end network. The joint formulation leads to an alternating minimization for deep feature representation at training stage and an efficient computation for heterogeneous data at testing stage. Extensive experiments on three challenging NIR-VIS face recognition databases demonstrate the significant superiority of Wasserstein CNN over state-of-the-art methods.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingOulu-CASIA NIR-VISTAR @ FAR=0.00154.6W-CNN He et al. (2018)
Facial Recognition and ModellingOulu-CASIA NIR-VISTAR @ FAR=0.0181.5W-CNN He et al. (2018)
Facial Recognition and ModellingBUAA-VisNirTAR @ FAR=0.00191.9W-CNN He et al. (2018)
Facial Recognition and ModellingBUAA-VisNirTAR @ FAR=0.0196W-CNN He et al. (2018)
Facial Recognition and ModellingCASIA NIR-VIS 2.0TAR @ FAR=0.00198.4W-CNN He et al. (2018)
Face VerificationOulu-CASIA NIR-VISTAR @ FAR=0.00154.6W-CNN He et al. (2018)
Face VerificationOulu-CASIA NIR-VISTAR @ FAR=0.0181.5W-CNN He et al. (2018)
Face VerificationBUAA-VisNirTAR @ FAR=0.00191.9W-CNN He et al. (2018)
Face VerificationBUAA-VisNirTAR @ FAR=0.0196W-CNN He et al. (2018)
Face VerificationCASIA NIR-VIS 2.0TAR @ FAR=0.00198.4W-CNN He et al. (2018)
Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.00154.6W-CNN He et al. (2018)
Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.0181.5W-CNN He et al. (2018)
Face ReconstructionBUAA-VisNirTAR @ FAR=0.00191.9W-CNN He et al. (2018)
Face ReconstructionBUAA-VisNirTAR @ FAR=0.0196W-CNN He et al. (2018)
Face ReconstructionCASIA NIR-VIS 2.0TAR @ FAR=0.00198.4W-CNN He et al. (2018)
3DOulu-CASIA NIR-VISTAR @ FAR=0.00154.6W-CNN He et al. (2018)
3DOulu-CASIA NIR-VISTAR @ FAR=0.0181.5W-CNN He et al. (2018)
3DBUAA-VisNirTAR @ FAR=0.00191.9W-CNN He et al. (2018)
3DBUAA-VisNirTAR @ FAR=0.0196W-CNN He et al. (2018)
3DCASIA NIR-VIS 2.0TAR @ FAR=0.00198.4W-CNN He et al. (2018)
3D Face ModellingOulu-CASIA NIR-VISTAR @ FAR=0.00154.6W-CNN He et al. (2018)
3D Face ModellingOulu-CASIA NIR-VISTAR @ FAR=0.0181.5W-CNN He et al. (2018)
3D Face ModellingBUAA-VisNirTAR @ FAR=0.00191.9W-CNN He et al. (2018)
3D Face ModellingBUAA-VisNirTAR @ FAR=0.0196W-CNN He et al. (2018)
3D Face ModellingCASIA NIR-VIS 2.0TAR @ FAR=0.00198.4W-CNN He et al. (2018)
3D Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.00154.6W-CNN He et al. (2018)
3D Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.0181.5W-CNN He et al. (2018)
3D Face ReconstructionBUAA-VisNirTAR @ FAR=0.00191.9W-CNN He et al. (2018)
3D Face ReconstructionBUAA-VisNirTAR @ FAR=0.0196W-CNN He et al. (2018)
3D Face ReconstructionCASIA NIR-VIS 2.0TAR @ FAR=0.00198.4W-CNN He et al. (2018)

Related Papers

ProxyFusion: Face Feature Aggregation Through Sparse Experts2025-09-24Non-Adaptive Adversarial Face Generation2025-07-16Attributes Shape the Embedding Space of Face Recognition Models2025-07-15Face mask detection project report.2025-07-02On the Burstiness of Faces in Set2025-06-25Identifying Physically Realizable Triggers for Backdoored Face Recognition Networks2025-06-24SELFI: Selective Fusion of Identity for Generalizable Deepfake Detection2025-06-21FaceLiVT: Face Recognition using Linear Vision Transformer with Structural Reparameterization For Mobile Device2025-06-12