TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Neural Architecture Search with Bayesian Optimisation and ...

Neural Architecture Search with Bayesian Optimisation and Optimal Transport

Kirthevasan Kandasamy, Willie Neiswanger, Jeff Schneider, Barnabas Poczos, Eric Xing

2018-02-11NeurIPS 2018 12Neural Architecture SearchModel SelectionBIG-bench Machine Learning
PaperPDFCode(official)

Abstract

Bayesian Optimisation (BO) refers to a class of methods for global optimisation of a function $f$ which is only accessible via point evaluations. It is typically used in settings where $f$ is expensive to evaluate. A common use case for BO in machine learning is model selection, where it is not possible to analytically model the generalisation performance of a statistical model, and we resort to noisy and expensive training and validation procedures to choose the best model. Conventional BO methods have focused on Euclidean and categorical domains, which, in the context of model selection, only permits tuning scalar hyper-parameters of machine learning algorithms. However, with the surge of interest in deep learning, there is an increasing demand to tune neural network \emph{architectures}. In this work, we develop NASBOT, a Gaussian process based BO framework for neural architecture search. To accomplish this, we develop a distance metric in the space of neural network architectures which can be computed efficiently via an optimal transport program. This distance might be of independent interest to the deep learning community as it may find applications outside of BO. We demonstrate that NASBOT outperforms other alternatives for architecture search in several cross validation based model selection tasks on multi-layer perceptrons and convolutional neural networks.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.37NASBOT
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Search time (s)75600NASBOT
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Test)46.37NASBOT
AutoMLNAS-Bench-201, ImageNet-16-120Search time (s)75600NASBOT

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17Topic Modeling and Link-Prediction for Material Property Discovery2025-07-08Advanced Financial Reasoning at Scale: A Comprehensive Evaluation of Large Language Models on CFA Level III2025-06-29mTSBench: Benchmarking Multivariate Time Series Anomaly Detection and Model Selection at Scale2025-06-26AnalogNAS-Bench: A NAS Benchmark for Analog In-Memory Computing2025-06-23From Tiny Machine Learning to Tiny Deep Learning: A Survey2025-06-21One-Shot Neural Architecture Search with Network Similarity Directed Initialization for Pathological Image Classification2025-06-17DDS-NAS: Dynamic Data Selection within Neural Architecture Search via On-line Hard Example Mining applied to Image Classification2025-06-17