TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CATE: Computation-aware Neural Architecture Encoding with ...

CATE: Computation-aware Neural Architecture Encoding with Transformers

Shen Yan, Kaiqiang Song, Fei Liu, Mi Zhang

2021-02-14Unsupervised Pre-trainingRepresentation LearningAutoMLNeural Architecture Search
PaperPDFCode(official)

Abstract

Recent works (White et al., 2020a; Yan et al., 2020) demonstrate the importance of architecture encodings in Neural Architecture Search (NAS). These encodings encode either structure or computation information of the neural architectures. Compared to structure-aware encodings, computation-aware encodings map architectures with similar accuracies to the same region, which improves the downstream architecture search performance (Zhang et al., 2019; White et al., 2020a). In this work, we introduce a Computation-Aware Transformer-based Encoding method called CATE. Different from existing computation-aware encodings based on fixed transformation (e.g. path encoding), CATE employs a pairwise pre-training scheme to learn computation-aware encodings using Transformers with cross-attention. Such learned encodings contain dense and contextualized computation information of neural architectures. We compare CATE with eleven encodings under three major encoding-dependent NAS subroutines in both small and large search spaces. Our experiments show that CATE is beneficial to the downstream search, especially in the large search space. Moreover, the outside search space experiment demonstrates its superior generalization ability beyond the search space on which it was trained. Our code is available at: https://github.com/MSU-MLSys-Lab/CATE.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchCIFAR-10 Image ClassificationParams4.1CATE
Neural Architecture SearchCIFAR-10 Image ClassificationPercentage error2.46CATE
Neural Architecture SearchCIFAR-10 Image ClassificationSearch Time (GPU days)10.3CATE
Neural Architecture SearchCIFAR-10Parameters4.1CATE
Neural Architecture SearchCIFAR-10Search Time (GPU days)10.3CATE
AutoMLCIFAR-10 Image ClassificationParams4.1CATE
AutoMLCIFAR-10 Image ClassificationPercentage error2.46CATE
AutoMLCIFAR-10 Image ClassificationSearch Time (GPU days)10.3CATE
AutoMLCIFAR-10Parameters4.1CATE
AutoMLCIFAR-10Search Time (GPU days)10.3CATE

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16Language-Guided Contrastive Audio-Visual Masked Autoencoder with Automatically Generated Audio-Visual-Text Triplets from Videos2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16