TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/GreedyNAS: Towards Fast One-Shot NAS with Greedy Supernet

GreedyNAS: Towards Fast One-Shot NAS with Greedy Supernet

Shan You, Tao Huang, Mingmin Yang, Fei Wang, Chen Qian, Chang-Shui Zhang

2020-03-25CVPR 2020 6Image ClassificationNeural Architecture SearchAll
PaperPDF

Abstract

Training a supernet matters for one-shot neural architecture search (NAS) methods since it serves as a basic performance estimator for different architectures (paths). Current methods mainly hold the assumption that a supernet should give a reasonable ranking over all paths. They thus treat all paths equally, and spare much effort to train paths. However, it is harsh for a single supernet to evaluate accurately on such a huge-scale search space (e.g., $7^{21}$). In this paper, instead of covering all paths, we ease the burden of supernet by encouraging it to focus more on evaluation of those potentially-good ones, which are identified using a surrogate portion of validation data. Concretely, during training, we propose a multi-path sampling strategy with rejection, and greedily filter the weak paths. The training efficiency is thus boosted since the training space has been greedily shrunk from all paths to those potentially-good ones. Moreover, we further adopt an exploration and exploitation policy by introducing an empirical candidate path pool. Our proposed method GreedyNAS is easy-to-follow, and experimental results on ImageNet dataset indicate that it can achieve better Top-1 accuracy under same search space and FLOPs or latency level, but with only $\sim$60\% of supernet training cost. By searching on a larger space, our GreedyNAS can also obtain new state-of-the-art architectures.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchImageNetAccuracy77.1GreedyNAS-A
Neural Architecture SearchImageNetTop-1 Error Rate22.9GreedyNAS-A
Neural Architecture SearchImageNetAccuracy76.8GreedyNAS-B
Neural Architecture SearchImageNetTop-1 Error Rate23.2GreedyNAS-B
Neural Architecture SearchImageNetAccuracy76.2GreedyNAS-C
Neural Architecture SearchImageNetTop-1 Error Rate23.8GreedyNAS-C
Image ClassificationImageNetGFLOPs0.366GreedyNAS-A
Image ClassificationImageNetGFLOPs0.324GreedyNAS-B
Image ClassificationImageNetGFLOPs0.284GreedyNAS-C
AutoMLImageNetAccuracy77.1GreedyNAS-A
AutoMLImageNetTop-1 Error Rate22.9GreedyNAS-A
AutoMLImageNetAccuracy76.8GreedyNAS-B
AutoMLImageNetTop-1 Error Rate23.2GreedyNAS-B
AutoMLImageNetAccuracy76.2GreedyNAS-C
AutoMLImageNetTop-1 Error Rate23.8GreedyNAS-C

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Modeling Code: Is Text All You Need?2025-07-15