TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/iDARTS: Differentiable Architecture Search with Stochastic...

iDARTS: Differentiable Architecture Search with Stochastic Implicit Gradients

Miao Zhang, Steven Su, Shirui Pan, Xiaojun Chang, Ehsan Abbasnejad, Reza Haffari

2021-06-21Neural Architecture Search
PaperPDFCode(official)

Abstract

\textit{Differentiable ARchiTecture Search} (DARTS) has recently become the mainstream of neural architecture search (NAS) due to its efficiency and simplicity. With a gradient-based bi-level optimization, DARTS alternately optimizes the inner model weights and the outer architecture parameter in a weight-sharing supernet. A key challenge to the scalability and quality of the learned architectures is the need for differentiating through the inner-loop optimisation. While much has been discussed about several potentially fatal factors in DARTS, the architecture gradient, a.k.a. hypergradient, has received less attention. In this paper, we tackle the hypergradient computation in DARTS based on the implicit function theorem, making it only depends on the obtained solution to the inner-loop optimization and agnostic to the optimization path. To further reduce the computational requirements, we formulate a stochastic hypergradient approximation for differentiable NAS, and theoretically show that the architecture optimization with the proposed method, named iDARTS, is expected to converge to a stationary point. Comprehensive experiments on two NAS benchmark search spaces and the common NAS search space verify the effectiveness of our proposed method. It leads to architectures outperforming, with large margins, those learned by the baseline methods.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)40.89iDARTS
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Val)40.38iDARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Test)93.58iDARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Val)89.86iDARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)70.83iDARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)70.57iDARTS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)40.89iDARTS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Val)40.38iDARTS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Test)93.58iDARTS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Val)89.86iDARTS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)70.83iDARTS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)70.57iDARTS

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17DDS-NAS: Dynamic Data Selection within Neural Architecture Search via On-line Hard Example Mining applied to Image Classification2025-06-17MARCO: Hardware-Aware Neural Architecture Search for Edge Devices with Multi-Agent Reinforcement Learning and Conformal Prediction Filtering2025-06-16Finding Optimal Kernel Size and Dimension in Convolutional Neural Networks An Architecture Optimization Approach2025-06-16Directed Acyclic Graph Convolutional Networks2025-06-13