TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TabTransformer: Tabular Data Modeling Using Contextual Emb...

TabTransformer: Tabular Data Modeling Using Contextual Embeddings

Xin Huang, Ashish Khetan, Milan Cvitkovic, Zohar Karnin

2020-12-11Unsupervised Pre-trainingtabular-classification
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCode

Abstract

We propose TabTransformer, a novel deep tabular data modeling architecture for supervised and semi-supervised learning. The TabTransformer is built upon self-attention based Transformers. The Transformer layers transform the embeddings of categorical features into robust contextual embeddings to achieve higher prediction accuracy. Through extensive experiments on fifteen publicly available datasets, we show that the TabTransformer outperforms the state-of-the-art deep learning methods for tabular data by at least 1.0% on mean AUC, and matches the performance of tree-based ensemble models. Furthermore, we demonstrate that the contextual embeddings learned from TabTransformer are highly robust against both missing and noisy data features, and provide better interpretability. Lastly, for the semi-supervised setting we develop an unsupervised pre-training procedure to learn data-driven contextual embeddings, resulting in an average 2.1% AUC lift over the state-of-the-art methods.

Related Papers

Qwen3 Embedding: Advancing Text Embedding and Reranking Through Foundation Models2025-06-05SPACE: Your Genomic Profile Predictor is a Powerful DNA Foundation Model2025-06-02Learning Interpretable Differentiable Logic Networks for Tabular Regression2025-05-29Foundation Model for Wireless Technology Recognition Using IQ Timeseries2025-05-26Is MixIT Really Unsuitable for Correlated Sources? Exploring MixIT for Unsupervised Pre-training in Music Source Separation2025-05-12The Efficiency of Pre-training with Objective Masking in Pseudo Labeling for Semi-Supervised Text Classification2025-05-10Latte: Transfering LLMs` Latent-level Knowledge for Few-shot Tabular Learning2025-05-08Risk Assessment Framework for Code LLMs via Leveraging Internal States2025-04-20