TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Does Unsupervised Architecture Representation Learning Hel...

Does Unsupervised Architecture Representation Learning Help Neural Architecture Search?

Shen Yan, Yu Zheng, Wei Ao, Xiao Zeng, Mi Zhang

2020-06-12NeurIPS 2020 12AutoMLNeural Architecture Search
PaperPDFCode(official)

Abstract

Existing Neural Architecture Search (NAS) methods either encode neural architectures using discrete encodings that do not scale well, or adopt supervised learning-based methods to jointly learn architecture representations and optimize architecture search on such representations which incurs search bias. Despite the widespread use, architecture representations learned in NAS are still poorly understood. We observe that the structural properties of neural architectures are hard to preserve in the latent space if architecture representation learning and search are coupled, resulting in less effective search performance. In this work, we find empirically that pre-training architecture representations using only neural architectures without their accuracies as labels considerably improve the downstream architecture search efficiency. To explain these observations, we visualize how unsupervised architecture representation learning better encourages neural architectures with similar connections and operators to cluster together. This helps to map neural architectures with similar performance to the same regions in the latent space and makes the transition of architectures in the latent space relatively smooth, which considerably benefits diverse downstream search strategies.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.27arch2vec
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Test)94.18arch2vec
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Val)91.41arch2vec
Neural Architecture SearchNAS-Bench-201, CIFAR-10Search time (s)12000arch2vec
Neural Architecture SearchCIFAR-10 Image ClassificationParams3.6arch2vec
Neural Architecture SearchCIFAR-10 Image ClassificationPercentage error2.56arch2vec
Neural Architecture SearchCIFAR-10 Image ClassificationSearch Time (GPU days)10.5arch2vec
Neural Architecture SearchCIFAR-10Search Time (GPU days)10.5arch2vec
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Test)73.37arch2vec
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)73.35arch2vec
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.27arch2vec
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Test)94.18arch2vec
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Val)91.41arch2vec
AutoMLNAS-Bench-201, CIFAR-10Search time (s)12000arch2vec
AutoMLCIFAR-10 Image ClassificationParams3.6arch2vec
AutoMLCIFAR-10 Image ClassificationPercentage error2.56arch2vec
AutoMLCIFAR-10 Image ClassificationSearch Time (GPU days)10.5arch2vec
AutoMLCIFAR-10Search Time (GPU days)10.5arch2vec
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Test)73.37arch2vec
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)73.35arch2vec

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17Imbalanced Regression Pipeline Recommendation2025-07-16Optimising 4th-Order Runge-Kutta Methods: A Dynamic Heuristic Approach for Efficiency and Low Storage2025-06-26Multimodal Representation Learning and Fusion2025-06-25Overtuning in Hyperparameter Optimization2025-06-24AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17