TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Scalable Cross-Entropy Loss for Sequential Recommendations...

Scalable Cross-Entropy Loss for Sequential Recommendations with Large Item Catalogs

Gleb Mezentsev, Danil Gusak, Ivan Oseledets, Evgeny Frolov

2024-09-27Sequential RecommendationRecommendation Systems
PaperPDFCode(official)

Abstract

Scalability issue plays a crucial role in productionizing modern recommender systems. Even lightweight architectures may suffer from high computational overload due to intermediate calculations, limiting their practicality in real-world applications. Specifically, applying full Cross-Entropy (CE) loss often yields state-of-the-art performance in terms of recommendations quality. Still, it suffers from excessive GPU memory utilization when dealing with large item catalogs. This paper introduces a novel Scalable Cross-Entropy (SCE) loss function in the sequential learning setup. It approximates the CE loss for datasets with large-size catalogs, enhancing both time efficiency and memory usage without compromising recommendations quality. Unlike traditional negative sampling methods, our approach utilizes a selective GPU-efficient computation strategy, focusing on the most informative elements of the catalog, particularly those most likely to be false positives. This is achieved by approximating the softmax distribution over a subset of the model outputs through the maximum inner product search. Experimental results on multiple datasets demonstrate the effectiveness of SCE in reducing peak memory usage by a factor of up to 100 compared to the alternatives, retaining or even exceeding their metrics values. The proposed approach also opens new perspectives for large-scale developments in different domains, such as large language models.

Results

TaskDatasetMetricValueModel
Recommendation SystemsAmazon BeautyHR@100.0935SASRec-SCE
Recommendation SystemsAmazon BeautyNDCG@100.0544SASRec-SCE
Recommendation SystemsGowallaCOV@10.0304SASRec-SCE
Recommendation SystemsGowallaCOV@100.219SASRec-SCE
Recommendation SystemsGowallaCOV@50.126SASRec-SCE
Recommendation SystemsGowallaHR@100.0831SASRec-SCE
Recommendation SystemsGowallaHR@50.0574SASRec-SCE
Recommendation SystemsGowallaNDCG@10.0207SASRec-SCE
Recommendation SystemsGowallaNDCG@100.0476SASRec-SCE
Recommendation SystemsGowallaNDCG@50.0393SASRec-SCE
Recommendation SystemsBehanceCOV@10.0393SASRec-SCE
Recommendation SystemsBehanceCOV@100.25SASRec-SCE
Recommendation SystemsBehanceCOV@515.3SASRec-SCE
Recommendation SystemsBehanceHR@100.113SASRec-SCE
Recommendation SystemsBehanceHR@50.0853SASRec-SCE
Recommendation SystemsBehanceNDCG@10.0277SASRec-SCE
Recommendation SystemsBehanceNDCG@100.0663SASRec-SCE
Recommendation SystemsBehanceNDCG@50.0572SASRec-SCE

Related Papers

IP2: Entity-Guided Interest Probing for Personalized News Recommendation2025-07-18A Reproducibility Study of Product-side Fairness in Bundle Recommendation2025-07-18SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16Looking for Fairness in Recommender Systems2025-07-16Journalism-Guided Agentic In-Context Learning for News Stance Detection2025-07-15LLM-Stackelberg Games: Conjectural Reasoning Equilibria and Their Applications to Spearphishing2025-07-12When Graph Contrastive Learning Backfires: Spectral Vulnerability and Defense in Recommendation2025-07-10