TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Distribution Matching for Multi-Task Learning of Classific...

Distribution Matching for Multi-Task Learning of Classification Tasks: a Large-Scale Study on Faces & Beyond

Dimitrios Kollias, Viktoriia Sharmanska, Stefanos Zafeiriou

2024-01-02Face RecognitionFacial Action Unit DetectionMulti-Task LearningFacial Expression Recognition (FER)Action Unit Detection
PaperPDF

Abstract

Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space, or parameter transfer. To provide sufficient learning support, modern MTL uses annotated data with full, or sufficiently large overlap across tasks, i.e., each input sample is annotated for all, or most of the tasks. However, collecting such annotations is prohibitive in many real applications, and cannot benefit from datasets available for individual tasks. In this work, we challenge this setup and show that MTL can be successful with classification tasks with little, or non-overlapping annotations, or when there is big discrepancy in the size of labeled data per task. We explore task-relatedness for co-annotation and co-training, and propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching. To demonstrate the general applicability of our method, we conducted diverse case studies in the domains of affective computing, face recognition, species recognition, and shopping item classification using nine datasets. Our large-scale study of affective tasks for basic expression recognition and facial action unit detection illustrates that our approach is network agnostic and brings large performance improvements compared to the state-of-the-art in both tasks and across all studied databases. In all case studies, we show that co-training via task-relatedness is advantageous and prevents negative transfer (which occurs when MT model's performance is worse than that of at least one single-task model).

Results

TaskDatasetMetricValueModel
Facial Recognition and ModellingRAF-DBAvg. Accuracy84.8C MT PSR
Facial Recognition and ModellingRAF-DBAvg. Accuracy81.4C MT VGGFACE
Face ReconstructionRAF-DBAvg. Accuracy84.8C MT PSR
Face ReconstructionRAF-DBAvg. Accuracy81.4C MT VGGFACE
Facial Expression Recognition (FER)RAF-DBAvg. Accuracy84.8C MT PSR
Facial Expression Recognition (FER)RAF-DBAvg. Accuracy81.4C MT VGGFACE
3DRAF-DBAvg. Accuracy84.8C MT PSR
3DRAF-DBAvg. Accuracy81.4C MT VGGFACE
3D Face ModellingRAF-DBAvg. Accuracy84.8C MT PSR
3D Face ModellingRAF-DBAvg. Accuracy81.4C MT VGGFACE
3D Face ReconstructionRAF-DBAvg. Accuracy84.8C MT PSR
3D Face ReconstructionRAF-DBAvg. Accuracy81.4C MT VGGFACE

Related Papers

ProxyFusion: Face Feature Aggregation Through Sparse Experts2025-09-24SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Non-Adaptive Adversarial Face Generation2025-07-16Attributes Shape the Embedding Space of Face Recognition Models2025-07-15Robust-Multi-Task Gradient Boosting2025-07-15SAMO: A Lightweight Sharpness-Aware Approach for Multi-Task Optimization with Joint Global-Local Perturbation2025-07-10Face mask detection project report.2025-07-02Multimodal Prompt Alignment for Facial Expression Recognition2025-06-26