TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/$Λ$-DARTS: Mitigating Performance Collapse by Harmonizing ...

$Λ$-DARTS: Mitigating Performance Collapse by Harmonizing Operation Selection among Cells

Sajad Movahedi, Melika Adabinejad, Ayyoob Imani, Arezou Keshavarz, Mostafa Dehghani, Azadeh Shakery, Babak N. Araabi

2022-10-14Neural Architecture Search
PaperPDFCode(official)

Abstract

Differentiable neural architecture search (DARTS) is a popular method for neural architecture search (NAS), which performs cell-search and utilizes continuous relaxation to improve the search efficiency via gradient-based optimization. The main shortcoming of DARTS is performance collapse, where the discovered architecture suffers from a pattern of declining quality during search. Performance collapse has become an important topic of research, with many methods trying to solve the issue through either regularization or fundamental changes to DARTS. However, the weight-sharing framework used for cell-search in DARTS and the convergence of architecture parameters has not been analyzed yet. In this paper, we provide a thorough and novel theoretical and empirical analysis on DARTS and its point of convergence. We show that DARTS suffers from a specific structural flaw due to its weight-sharing framework that limits the convergence of DARTS to saturation points of the softmax function. This point of convergence gives an unfair advantage to layers closer to the output in choosing the optimal architecture, causing performance collapse. We then propose two new regularization terms that aim to prevent performance collapse by harmonizing operation selection via aligning gradients of layers. Experimental results on six different search spaces and three different datasets show that our method ($\Lambda$-DARTS) does indeed prevent performance collapse, providing justification for our theoretical analysis and the proposed remedy.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.34Λ-DARTS
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Val)46.37Λ-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Test)94.36Λ-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Val)91.55Λ-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)73.51Λ-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)73.49Λ-DARTS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.34Λ-DARTS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Val)46.37Λ-DARTS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Test)94.36Λ-DARTS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Val)91.55Λ-DARTS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)73.51Λ-DARTS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)73.49Λ-DARTS

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17DDS-NAS: Dynamic Data Selection within Neural Architecture Search via On-line Hard Example Mining applied to Image Classification2025-06-17MARCO: Hardware-Aware Neural Architecture Search for Edge Devices with Multi-Agent Reinforcement Learning and Conformal Prediction Filtering2025-06-16Finding Optimal Kernel Size and Dimension in Convolutional Neural Networks An Architecture Optimization Approach2025-06-16Directed Acyclic Graph Convolutional Networks2025-06-13