TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Transformer-Based Auxiliary Loss for Face Recognition Acro...

Transformer-Based Auxiliary Loss for Face Recognition Across Age Variations

Pritesh Prakash, Ashish Jacob Sam, S Umamaheswaran

2024-12-03Face Recognition
PaperPDF

Abstract

Aging presents a significant challenge in face recognition, as changes in skin texture and tone can alter facial features over time, making it particularly difficult to compare images of the same individual taken years apart, such as in long-term identification scenarios. Transformer networks have the strength to preserve sequential spatial relationships caused by aging effect. This paper presents a technique for loss evaluation that uses a transformer network as an additive loss in the face recognition domain. The standard metric loss function typically takes the final embedding of the main CNN backbone as its input. Here, we employ a transformer-metric loss, a combined approach that integrates both transformer-loss and metric-loss. This research intends to analyze the transformer behavior on the convolution output when the CNN outcome is arranged in a sequential vector. These sequential vectors have the potential to overcome the texture or regional structure referred to as wrinkles or sagging skin affected by aging. The transformer encoder takes input from the contextual vectors obtained from the final convolution layer of the network. The learned features can be more age-invariant, complementing the discriminative power of the standard metric loss embedding. With this technique, we use transformer loss with various base metric-loss functions to evaluate the effect of the combined loss functions. We observe that such a configuration allows the network to achieve SoTA results in LFW and age-variant datasets (CA-LFW and AgeDB). This research expands the role of transformers in the machine vision domain and opens new possibilities for exploring transformers as a loss function.

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingAgeDB-30Accuracy0.9831Transformer loss+ArcFace ResNet100
Face ReconstructionAgeDB-30Accuracy0.9831Transformer loss+ArcFace ResNet100
Face RecognitionAgeDB-30Accuracy0.9831Transformer loss+ArcFace ResNet100
3DAgeDB-30Accuracy0.9831Transformer loss+ArcFace ResNet100
3D Face ModellingAgeDB-30Accuracy0.9831Transformer loss+ArcFace ResNet100
3D Face ReconstructionAgeDB-30Accuracy0.9831Transformer loss+ArcFace ResNet100

Related Papers

ProxyFusion: Face Feature Aggregation Through Sparse Experts2025-09-24Non-Adaptive Adversarial Face Generation2025-07-16Attributes Shape the Embedding Space of Face Recognition Models2025-07-15Face mask detection project report.2025-07-02On the Burstiness of Faces in Set2025-06-25Identifying Physically Realizable Triggers for Backdoored Face Recognition Networks2025-06-24SELFI: Selective Fusion of Identity for Generalizable Deepfake Detection2025-06-21FaceLiVT: Face Recognition using Linear Vision Transformer with Structural Reparameterization For Mobile Device2025-06-12