TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Neural Factorization Machines for Sparse Predictive Analyt...

Neural Factorization Machines for Sparse Predictive Analytics

Xiangnan He, Tat-Seng Chua

2017-08-16Link Prediction
PaperPDFCode(official)CodeCodeCodeCodeCode

Abstract

Many predictive tasks of web applications need to model categorical variables, such as user IDs and demographics like genders and occupations. To apply standard machine learning techniques, these categorical predictors are always converted to a set of binary features via one-hot encoding, making the resultant feature vector highly sparse. To learn from such sparse data effectively, it is crucial to account for the interactions between features. Factorization Machines (FMs) are a popular solution for efficiently using the second-order feature interactions. However, FM models feature interactions in a linear way, which can be insufficient for capturing the non-linear and complex inherent structure of real-world data. While deep neural networks have recently been applied to learn non-linear feature interactions in industry, such as the Wide&Deep by Google and DeepCross by Microsoft, the deep structure meanwhile makes them difficult to train. In this paper, we propose a novel model Neural Factorization Machine (NFM) for prediction under sparse settings. NFM seamlessly combines the linearity of FM in modelling second-order feature interactions and the non-linearity of neural network in modelling higher-order feature interactions. Conceptually, NFM is more expressive than FM since FM can be seen as a special case of NFM without hidden layers. Empirical results on two regression tasks show that with one hidden layer only, NFM significantly outperforms FM with a 7.3% relative improvement. Compared to the recent deep learning methods Wide&Deep and DeepCross, our NFM uses a shallower structure but offers better performance, being much easier to train and tune in practice.

Results

TaskDatasetMetricValueModel
Link PredictionMovieLens 25MHits@100.8132NFM
Link PredictionMovieLens 25MnDCG@100.5347NFM
Link PredictionYelpHR@100.8595NFM
Link PredictionYelpnDCG@100.6062NFM

Related Papers

Topic Modeling and Link-Prediction for Material Property Discovery2025-07-08Graph Collaborative Attention Network for Link Prediction in Knowledge Graphs2025-07-05Understanding Generalization in Node and Link Prediction2025-07-01Context-Driven Knowledge Graph Completion with Semantic-Aware Relational Message Passing2025-06-29Directed Link Prediction using GNN with Local and Global Feature Fusion2025-06-25Demystifying Distributed Training of Graph Neural Networks for Link Prediction2025-06-25Call Me Maybe: Enhancing JavaScript Call Graph Construction using Graph Neural Networks2025-06-22A Transfer Learning Framework for Multilayer Networks via Model Averaging2025-06-14