TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/$β$-DARTS: Beta-Decay Regularization for Differentiable Ar...

$β$-DARTS: Beta-Decay Regularization for Differentiable Architecture Search

Peng Ye, Baopu Li, Yikang Li, Tao Chen, Jiayuan Fan, Wanli Ouyang

2022-03-03Neural Architecture Search
PaperPDFCode(official)

Abstract

Neural Architecture Search~(NAS) has attracted increasingly more attention in recent years because of its capability to design deep neural networks automatically. Among them, differential NAS approaches such as DARTS, have gained popularity for the search efficiency. However, they suffer from two main issues, the weak robustness to the performance collapse and the poor generalization ability of the searched architectures. To solve these two problems, a simple-but-efficient regularization method, termed as Beta-Decay, is proposed to regularize the DARTS-based NAS searching process. Specifically, Beta-Decay regularization can impose constraints to keep the value and variance of activated architecture parameters from too large. Furthermore, we provide in-depth theoretical analysis on how it works and why it works. Experimental results on NAS-Bench-201 show that our proposed method can help to stabilize the searching process and makes the searched network more transferable across different datasets. In addition, our search scheme shows an outstanding property of being less dependent on training time and data. Comprehensive experiments on a variety of search spaces and datasets validate the effectiveness of the proposed method.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.71β-SDARTS-RS
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.71β-RDARTS-L2
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.34β-DARTS
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Val)46.37β-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Test)94.36β-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Val)91.55β-DARTS
Neural Architecture SearchCIFAR-100Percentage Error16.52β-DARTS
Neural Architecture SearchImageNetTop-1 Error Rate23.9b-DARTS (CIFAR-10)
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)73.51β-DARTS
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)73.49β-DARTS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.71β-SDARTS-RS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.71β-RDARTS-L2
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.34β-DARTS
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Val)46.37β-DARTS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Test)94.36β-DARTS
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Val)91.55β-DARTS
AutoMLCIFAR-100Percentage Error16.52β-DARTS
AutoMLImageNetTop-1 Error Rate23.9b-DARTS (CIFAR-10)
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)73.51β-DARTS
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)73.49β-DARTS

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17DDS-NAS: Dynamic Data Selection within Neural Architecture Search via On-line Hard Example Mining applied to Image Classification2025-06-17MARCO: Hardware-Aware Neural Architecture Search for Edge Devices with Multi-Agent Reinforcement Learning and Conformal Prediction Filtering2025-06-16Finding Optimal Kernel Size and Dimension in Convolutional Neural Networks An Architecture Optimization Approach2025-06-16Directed Acyclic Graph Convolutional Networks2025-06-13