TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Disentangled Variational Representation for Heterogeneous ...

Disentangled Variational Representation for Heterogeneous Face Recognition

Xiang Wu, Huaibo Huang, Vishal M. Patel, Ran He, Zhenan Sun

2018-09-06Face RecognitionHeterogeneous Face Recognition
PaperPDF

Abstract

Visible (VIS) to near infrared (NIR) face matching is a challenging problem due to the significant domain discrepancy between the domains and a lack of sufficient data for training cross-modal matching algorithms. Existing approaches attempt to tackle this problem by either synthesizing visible faces from NIR faces, extracting domain-invariant features from these modalities, or projecting heterogeneous data onto a common latent space for cross-modal matching. In this paper, we take a different approach in which we make use of the Disentangled Variational Representation (DVR) for cross-modal matching. First, we model a face representation with an intrinsic identity information and its within-person variations. By exploring the disentangled latent variable space, a variational lower bound is employed to optimize the approximate posterior for NIR and VIS representations. Second, aiming at obtaining more compact and discriminative disentangled latent space, we impose a minimization of the identity information for the same subject and a relaxed correlation alignment constraint between the NIR and VIS modality variations. An alternative optimization scheme is proposed for the disentangled variational representation part and the heterogeneous face recognition network part. The mutual promotion between these two parts effectively reduces the NIR and VIS domain discrepancy and alleviates over-fitting. Extensive experiments on three challenging NIR-VIS heterogeneous face recognition databases demonstrate that the proposed method achieves significant improvements over the state-of-the-art methods.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingOulu-CASIA NIR-VISTAR @ FAR=0.00184.9DVR Wu et al. (2019)
Facial Recognition and ModellingOulu-CASIA NIR-VISTAR @ FAR=0.0197.2DVR Wu et al. (2019)
Facial Recognition and ModellingBUAA-VisNirTAR @ FAR=0.00196.9DVR Wu et al. (2019)
Facial Recognition and ModellingBUAA-VisNirTAR @ FAR=0.0198.5DVR Wu et al. (2019)
Facial Recognition and ModellingCASIA NIR-VIS 2.0TAR @ FAR=0.00199.6DVR Wu et al. (2019)
Face VerificationOulu-CASIA NIR-VISTAR @ FAR=0.00184.9DVR Wu et al. (2019)
Face VerificationOulu-CASIA NIR-VISTAR @ FAR=0.0197.2DVR Wu et al. (2019)
Face VerificationBUAA-VisNirTAR @ FAR=0.00196.9DVR Wu et al. (2019)
Face VerificationBUAA-VisNirTAR @ FAR=0.0198.5DVR Wu et al. (2019)
Face VerificationCASIA NIR-VIS 2.0TAR @ FAR=0.00199.6DVR Wu et al. (2019)
Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.00184.9DVR Wu et al. (2019)
Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.0197.2DVR Wu et al. (2019)
Face ReconstructionBUAA-VisNirTAR @ FAR=0.00196.9DVR Wu et al. (2019)
Face ReconstructionBUAA-VisNirTAR @ FAR=0.0198.5DVR Wu et al. (2019)
Face ReconstructionCASIA NIR-VIS 2.0TAR @ FAR=0.00199.6DVR Wu et al. (2019)
3DOulu-CASIA NIR-VISTAR @ FAR=0.00184.9DVR Wu et al. (2019)
3DOulu-CASIA NIR-VISTAR @ FAR=0.0197.2DVR Wu et al. (2019)
3DBUAA-VisNirTAR @ FAR=0.00196.9DVR Wu et al. (2019)
3DBUAA-VisNirTAR @ FAR=0.0198.5DVR Wu et al. (2019)
3DCASIA NIR-VIS 2.0TAR @ FAR=0.00199.6DVR Wu et al. (2019)
3D Face ModellingOulu-CASIA NIR-VISTAR @ FAR=0.00184.9DVR Wu et al. (2019)
3D Face ModellingOulu-CASIA NIR-VISTAR @ FAR=0.0197.2DVR Wu et al. (2019)
3D Face ModellingBUAA-VisNirTAR @ FAR=0.00196.9DVR Wu et al. (2019)
3D Face ModellingBUAA-VisNirTAR @ FAR=0.0198.5DVR Wu et al. (2019)
3D Face ModellingCASIA NIR-VIS 2.0TAR @ FAR=0.00199.6DVR Wu et al. (2019)
3D Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.00184.9DVR Wu et al. (2019)
3D Face ReconstructionOulu-CASIA NIR-VISTAR @ FAR=0.0197.2DVR Wu et al. (2019)
3D Face ReconstructionBUAA-VisNirTAR @ FAR=0.00196.9DVR Wu et al. (2019)
3D Face ReconstructionBUAA-VisNirTAR @ FAR=0.0198.5DVR Wu et al. (2019)
3D Face ReconstructionCASIA NIR-VIS 2.0TAR @ FAR=0.00199.6DVR Wu et al. (2019)

Related Papers

ProxyFusion: Face Feature Aggregation Through Sparse Experts2025-09-24Non-Adaptive Adversarial Face Generation2025-07-16Attributes Shape the Embedding Space of Face Recognition Models2025-07-15Face mask detection project report.2025-07-02On the Burstiness of Faces in Set2025-06-25Identifying Physically Realizable Triggers for Backdoored Face Recognition Networks2025-06-24SELFI: Selective Fusion of Identity for Generalizable Deepfake Detection2025-06-21FaceLiVT: Face Recognition using Linear Vision Transformer with Structural Reparameterization For Mobile Device2025-06-12