TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SNAS: Stochastic Neural Architecture Search

SNAS: Stochastic Neural Architecture Search

Sirui Xie, Hehui Zheng, Chunxiao Liu, Liang Lin

2018-12-24ICLR 2019 5Reinforcement LearningNeural Architecture Searchreinforcement-learning
PaperPDFCodeCode(official)

Abstract

We propose Stochastic Neural Architecture Search (SNAS), an economical end-to-end solution to Neural Architecture Search (NAS) that trains neural operation parameters and architecture distribution parameters in same round of back-propagation, while maintaining the completeness and differentiability of the NAS pipeline. In this work, NAS is reformulated as an optimization problem on parameters of a joint distribution for the search space in a cell. To leverage the gradient information in generic differentiable loss for architecture search, a novel search gradient is proposed. We prove that this search gradient optimizes the same objective as reinforcement-learning-based NAS, but assigns credits to structural decisions more efficiently. This credit assignment is further augmented with locally decomposable reward to enforce a resource-efficient constraint. In experiments on CIFAR-10, SNAS takes less epochs to find a cell architecture with state-of-the-art accuracy than non-differentiable evolution-based and reinforcement-learning-based NAS, which is also transferable to ImageNet. It is also shown that child networks of SNAS can maintain the validation accuracy in searching, with which attention-based NAS requires parameter retraining to compete, exhibiting potentials to stride towards efficient NAS on big datasets. We have released our implementation at https://github.com/SNAS-Series/SNAS-Series.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)43.16SNAS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Test)92.77SNAS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Val)90.1SNAS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)69.34SNAS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)69.69SNAS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)43.16SNAS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Test)92.77SNAS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Val)90.1SNAS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)69.34SNAS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)69.69SNAS

Related Papers

CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning2025-07-18VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Aligning Humans and Robots via Reinforcement Learning from Implicit Human Feedback2025-07-17VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Autonomous Resource Management in Microservice Systems via Reinforcement Learning2025-07-17