TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A unifying mutual information view of metric learning: cro...

A unifying mutual information view of metric learning: cross-entropy vs. pairwise losses

Malik Boudiaf, Jérôme Rony, Imtiaz Masud Ziko, Eric Granger, Marco Pedersoli, Pablo Piantanida, Ismail Ben Ayed

2020-03-19ECCV 2020 8Metric Learning
PaperPDFCode(official)

Abstract

Recently, substantial research efforts in Deep Metric Learning (DML) focused on designing complex pairwise-distance losses, which require convoluted schemes to ease optimization, such as sample mining or pair weighting. The standard cross-entropy loss for classification has been largely overlooked in DML. On the surface, the cross-entropy may seem unrelated and irrelevant to metric learning as it does not explicitly involve pairwise distances. However, we provide a theoretical analysis that links the cross-entropy to several well-known and recent pairwise losses. Our connections are drawn from two different perspectives: one based on an explicit optimization insight; the other on discriminative and generative views of the mutual information between the labels and the learned features. First, we explicitly demonstrate that the cross-entropy is an upper bound on a new pairwise loss, which has a structure similar to various pairwise losses: it minimizes intra-class distances while maximizing inter-class distances. As a result, minimizing the cross-entropy can be seen as an approximate bound-optimization (or Majorize-Minimize) algorithm for minimizing this pairwise loss. Second, we show that, more generally, minimizing the cross-entropy is actually equivalent to maximizing the mutual information, to which we connect several well-known pairwise losses. Furthermore, we show that various standard pairwise losses can be explicitly related to one another via bound relationships. Our findings indicate that the cross-entropy represents a proxy for maximizing the mutual information -- as pairwise losses do -- without the need for convoluted sample-mining heuristics. Our experiments over four standard DML benchmarks strongly support our findings. We obtain state-of-the-art results, outperforming recent and complex DML methods.

Results

TaskDatasetMetricValueModel
Metric LearningCARS196R@189.3ResNet-50 + Cross-Entropy
Metric Learning CUB-200-2011R@169.2ResNet-50 + Cross-Entropy
Metric LearningIn-ShopR@190.6ResNet-50 + Cross-Entropy
Metric LearningStanford Online ProductsR@181.1ResNet-50 + Cross-Entropy

Related Papers

Unsupervised Ground Metric Learning2025-07-17Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16$\texttt{Droid}$: A Resource Suite for AI-Generated Code Detection2025-07-11Grid-Reg: Grid-Based SAR and Optical Image Registration Across Platforms2025-07-06Dare to Plagiarize? Plagiarized Painting Recognition and Retrieval2025-06-29Multimodal Information Retrieval for Open World with Edit Distance Weak Supervision2025-06-25Offline Goal-Conditioned Reinforcement Learning with Projective Quasimetric Planning2025-06-23AbRank: A Benchmark Dataset and Metric-Learning Framework for Antibody-Antigen Affinity Ranking2025-06-21