TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/An Effective Baseline for Robustness to Distributional Shift

An Effective Baseline for Robustness to Distributional Shift

Sunil Thulasidasan, Sushil Thapa, Sayera Dhaubhadel, Gopinath Chennupati, Tanmoy Bhattacharya, Jeff Bilmes

2021-05-15Robust classificationOut-of-Distribution Detectiontext-classificationDeep Learning
PaperPDFCode(official)

Abstract

Refraining from confidently predicting when faced with categories of inputs different from those seen during training is an important requirement for the safe deployment of deep learning systems. While simple to state, this has been a particularly challenging problem in deep learning, where models often end up making overconfident predictions in such situations. In this work we present a simple, but highly effective approach to deal with out-of-distribution detection that uses the principle of abstention: when encountering a sample from an unseen class, the desired behavior is to abstain from predicting. Our approach uses a network with an extra abstention class and is trained on a dataset that is augmented with an uncurated set that consists of a large number of out-of-distribution (OoD) samples that are assigned the label of the abstention class; the model is then trained to learn an effective discriminator between in and out-of-distribution samples. We compare this relatively simple approach against a wide variety of more complex methods that have been proposed both for out-of-distribution detection as well as uncertainty modeling in deep learning, and empirically demonstrate its effectiveness on a wide variety of of benchmarks and deep architectures for image recognition and text classification, often outperforming existing approaches by significant margins. Given the simplicity and effectiveness of this method, we propose that this approach be used as a new additional baseline for future work in this domain.

Results

TaskDatasetMetricValueModel
Out-of-Distribution DetectionTREC-NEWSAUROC99.92-Layered GRU
Out-of-Distribution DetectionTREC-NEWSFPR954.72-Layered GRU
Out-of-Distribution DetectionCIFAR-100AUROC97.7Wide ResNet 40x2
Out-of-Distribution DetectionCIFAR-100FPR9523.4Wide ResNet 40x2
Out-of-Distribution DetectionSSTAUROC99.72-Layered GRU
Out-of-Distribution DetectionSSTFPR9520.92-Layered GRU
Out-of-Distribution Detection20 NewsgroupsAUROC99.62-Layered GRU
Out-of-Distribution Detection20 NewsgroupsFPR951.782-Layered GRU
Out-of-Distribution DetectionCIFAR-10AUROC99.9Wide ResNet 40x2
Out-of-Distribution DetectionCIFAR-10FPR952Wide ResNet 40x2

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Making Language Model a Hierarchical Classifier and Generator2025-07-17A Survey of Deep Learning for Geometry Problem Solving2025-07-16GNN-CNN: An Efficient Hybrid Model of Convolutional and Graph Neural Networks for Text Representation2025-07-10Uncertainty Quantification for Motor Imagery BCI -- Machine Learning vs. Deep Learning2025-07-10Safe Domain Randomization via Uncertainty-Aware Out-of-Distribution Detection and Policy Adaptation2025-07-08Chat-Ghosting: A Comparative Study of Methods for Auto-Completion in Dialog Systems2025-07-08Deep Learning Optimization of Two-State Pinching Antennas Systems2025-07-08