TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CheXpert: A Large Chest Radiograph Dataset with Uncertaint...

CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison

Jeremy Irvin, Pranav Rajpurkar, Michael Ko, Yifan Yu, Silviana Ciurea-Ilcus, Chris Chute, Henrik Marklund, Behzad Haghgoo, Robyn Ball, Katie Shpanskaya, Jayne Seekins, David A. Mong, Safwan S. Halabi, Jesse K. Sandberg, Ricky Jones, David B. Larson, Curtis P. Langlotz, Bhavik N. Patel, Matthew P. Lungren, Andrew Y. Ng

2019-01-21Lung Disease Classification
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Large, labeled datasets have driven deep learning methods to achieve expert-level performance on a variety of medical imaging tasks. We present CheXpert, a large dataset that contains 224,316 chest radiographs of 65,240 patients. We design a labeler to automatically detect the presence of 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation. We investigate different approaches to using the uncertainty labels for training convolutional neural networks that output the probability of these observations given the available frontal and lateral radiographs. On a validation set of 200 chest radiographic studies which were manually annotated by 3 board-certified radiologists, we find that different uncertainty approaches are useful for different pathologies. We then evaluate our best model on a test set composed of 500 chest radiographic studies annotated by a consensus of 5 board-certified radiologists, and compare the performance of our model to that of 3 additional radiologists in the detection of 5 selected pathologies. On Cardiomegaly, Edema, and Pleural Effusion, the model ROC and PR curves lie above all 3 radiologist operating points. We release the dataset to the public as a standard benchmark to evaluate performance of chest radiograph interpretation models. The dataset is freely available at https://stanfordmlgroup.github.io/competitions/chexpert .

Results

TaskDatasetMetricValueModel
Multi-Label ClassificationCheXpertAVERAGE AUC ON 14 LABEL0.907Stanford Baseline (ensemble)
Multi-Label ClassificationCheXpertNUM RADS BELOW CURVE1.8Stanford Baseline (ensemble)

Related Papers

CXR-LT 2024: A MICCAI challenge on long-tailed, multi-label, and zero-shot disease classification from chest X-ray2025-06-09Multilabel Classification for Lung Disease Detection: Integrating Deep Learning and Natural Language Processing2024-12-16Developing a Dual-Stage Vision Transformer Model for Lung Disease Classification2024-09-26Deep Learning for Lung Disease Classification Using Transfer Learning and a Customized CNN Architecture with Attention2024-08-23ILDNet: A Novel Deep Learning Framework for Interstitial Lung Disease Identification Using Respiratory Sounds2024-08-22CROCODILE: Causality aids RObustness via COntrastive DIsentangled LEarning2024-08-09Non-contact Lung Disease Classification via OFDM-based Passive 6G ISAC Sensing2024-05-15Explainable Lung Disease Classification from Chest X-Ray Images Utilizing Deep Learning and XAI2024-04-17