TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/TABPFN

TABPFN

tabular data Prior-data Fitted Network

GeneralIntroduced 200026 papers
Source Paper

Description

We present TabPFN, a trained Transformer that can do supervised classification for small tabular datasets in less than a second, needs no hyperparameter tuning and is competitive with state-of-the-art classification methods. TabPFN is fully entailed in the weights of our network, which accepts training and test samples as a set-valued input and yields predictions for the entire test set in a single forward pass. TabPFN is a Prior-Data Fitted Network (PFN) and is trained offline once, to approximate Bayesian inference on synthetic datasets drawn from our prior. This prior incorporates ideas from causal reasoning: It entails a large space of structural causal models with a preference for simple structures. On the 18 datasets in the OpenML-CC18 suite that contain up to 1 000 training data points, up to 100 purely numerical features without missing values, and up to 10 classes, we show that our method clearly outperforms boosted trees and performs on par with complex state-of-the-art AutoML systems with up to 230× speedup. This increases to a 5 700× speedup when using a GPU. We also validate these results on an additional 67 small numerical datasets from OpenML. We provide all our code, the trained TabPFN, an interactive browser demo and a Colab notebook at https://github.com/automl/TabPFN.

Papers Using This Method

ConTextTab: A Semantics-Aware Tabular In-Context Learner2025-06-12On the Robustness of Tabular Foundation Models: Test-Time Attacks and In-Context Defenses2025-06-03TabPFN: One Model to Rule Them All?2025-05-26Realistic Evaluation of TabPFN v2 in Open Environments2025-05-22Tabular foundation model to detect empathy from visual cues2025-04-15A Closer Look at TabPFN v2: Strength, Limitation, and Extension2025-02-24TabPFN Unleashed: A Scalable and Effective Solution to Tabular Classification Problems2025-02-04Transformers Boost the Performance of Decision Trees on Tabular Data across Sample Sizes2025-02-04The Tabular Foundation Model TabPFN Outperforms Specialized Time Series Forecasting Models Based on Simple Features2025-01-06Drift-Resilient TabPFN: In-Context Learning Temporal Distribution Shifts on Tabular Data2024-11-15A Survey on Deep Tabular Learning2024-10-15AnnotatedTables: A Large Tabular Dataset with Language Model Annotations2024-06-24Large Scale Transfer Learning for Tabular Data via Language Modeling2024-06-17Tokenize features, enhancing tables: the FT-TABPFN model for tabular classification2024-06-11Grapevine Disease Prediction Using Climate Variables from Multi-Sensor Remote Sensing Imagery via a Transformer Model2024-06-11TabPFGen -- Tabular Data Generation with TabPFN2024-06-07Retrieval & Fine-Tuning for In-Context Tabular Models2024-06-07Fine-tuned In-Context Learning Transformers are Excellent Tabular Data Classifiers2024-05-22Interpretable Machine Learning for TabPFN2024-03-16TuneTables: Context Optimization for Scalable Prior-Data Fitted Networks2024-02-17