TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Geometry-Aware Gradient Algorithms for Neural Architecture...

Geometry-Aware Gradient Algorithms for Neural Architecture Search

Liam Li, Mikhail Khodak, Maria-Florina Balcan, Ameet Talwalkar

2020-04-16ICLR 2021 1Neural Architecture Search
PaperPDFCode(official)

Abstract

Recent state-of-the-art methods for neural architecture search (NAS) exploit gradient-based optimization by relaxing the problem into continuous optimization over architectures and shared-weights, a noisy process that remains poorly understood. We argue for the study of single-level empirical risk minimization to understand NAS with weight-sharing, reducing the design of NAS methods to devising optimizers and regularizers that can quickly obtain high-quality solutions to this problem. Invoking the theory of mirror descent, we present a geometry-aware framework that exploits the underlying structure of this optimization to return sparse architectural parameters, leading to simple yet novel algorithms that enjoy fast convergence guarantees and achieve state-of-the-art accuracy on the latest NAS benchmarks in computer vision. Notably, we exceed the best published results for both CIFAR and ImageNet on both the DARTS search space and NAS-Bench201; on the latter we achieve near-oracle-optimal performance on CIFAR-10 and CIFAR-100. Together, our theory and experiments demonstrate a principled way to co-design optimizers and continuous relaxations of discrete NAS search spaces.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.36GAEA DARTS (ERM)
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Test)94.1GAEA DARTS (ERM)
Neural Architecture SearchImageNetParams5.6GAEA PC-DARTS
Neural Architecture SearchImageNetTop-1 Error Rate24GAEA PC-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)73.43GAEA DARTS (ERM)
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.36GAEA DARTS (ERM)
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Test)94.1GAEA DARTS (ERM)
AutoMLImageNetParams5.6GAEA PC-DARTS
AutoMLImageNetTop-1 Error Rate24GAEA PC-DARTS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)73.43GAEA DARTS (ERM)

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17DDS-NAS: Dynamic Data Selection within Neural Architecture Search via On-line Hard Example Mining applied to Image Classification2025-06-17MARCO: Hardware-Aware Neural Architecture Search for Edge Devices with Multi-Agent Reinforcement Learning and Conformal Prediction Filtering2025-06-16Finding Optimal Kernel Size and Dimension in Convolutional Neural Networks An Architecture Optimization Approach2025-06-16Directed Acyclic Graph Convolutional Networks2025-06-13