TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/One-Shot Neural Architecture Search via Self-Evaluated Tem...

One-Shot Neural Architecture Search via Self-Evaluated Template Network

Xuanyi Dong, Yi Yang

2019-10-13ICCV 2019 10Neural Architecture Search
PaperPDFCodeCode(official)CodeCode(official)

Abstract

Neural architecture search (NAS) aims to automate the search procedure of architecture instead of manual design. Even if recent NAS approaches finish the search within days, lengthy training is still required for a specific architecture candidate to get the parameters for its accurate evaluation. Recently one-shot NAS methods are proposed to largely squeeze the tedious training process by sharing parameters across candidates. In this way, the parameters for each candidate can be directly extracted from the shared parameters instead of training them from scratch. However, they have no sense of which candidate will perform better until evaluation so that the candidates to evaluate are randomly sampled and the top-1 candidate is considered the best. In this paper, we propose a Self-Evaluated Template Network (SETN) to improve the quality of the architecture candidates for evaluation so that it is more likely to cover competitive candidates. SETN consists of two components: (1) an evaluator, which learns to indicate the probability of each individual architecture being likely to have a lower validation loss. The candidates for evaluation can thus be selectively sampled according to this evaluator. (2) a template network, which shares parameters among all candidates to amortize the training cost of generated candidates. In experiments, the architecture found by SETN achieves state-of-the-art performance on CIFAR and ImageNet benchmarks within comparable computation costs. Code is publicly available on GitHub: https://github.com/D-X-Y/AutoDL-Projects.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Val)32.52SETN
Neural Architecture SearchNAS-Bench-201, CIFAR-10Search time (s)31010SETN
Neural Architecture SearchCIFAR-10Search Time (GPU days)1.8SETN (T=1K) + CutOut
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)56.87SETN
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)59.05SETN
Neural Architecture SearchNAS-Bench-201, CIFAR-100Search time (s)31010SETN
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Val)32.52SETN
AutoMLNAS-Bench-201, CIFAR-10Search time (s)31010SETN
AutoMLCIFAR-10Search Time (GPU days)1.8SETN (T=1K) + CutOut
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)56.87SETN
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)59.05SETN
AutoMLNAS-Bench-201, CIFAR-100Search time (s)31010SETN

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17DDS-NAS: Dynamic Data Selection within Neural Architecture Search via On-line Hard Example Mining applied to Image Classification2025-06-17MARCO: Hardware-Aware Neural Architecture Search for Edge Devices with Multi-Agent Reinforcement Learning and Conformal Prediction Filtering2025-06-16Finding Optimal Kernel Size and Dimension in Convolutional Neural Networks An Architecture Optimization Approach2025-06-16Directed Acyclic Graph Convolutional Networks2025-06-13