TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CATCH: Context-based Meta Reinforcement Learning for Trans...

CATCH: Context-based Meta Reinforcement Learning for Transferrable Architecture Search

Xin Chen, Yawen Duan, Zewei Chen, Hang Xu, Zihao Chen, Xiaodan Liang, Tong Zhang, Zhenguo Li

2020-07-18ECCV 2020 8Meta-LearningReinforcement LearningNeural Architecture SearchMeta Reinforcement Learningreinforcement-learning
PaperPDF

Abstract

Neural Architecture Search (NAS) achieved many breakthroughs in recent years. In spite of its remarkable progress, many algorithms are restricted to particular search spaces. They also lack efficient mechanisms to reuse knowledge when confronting multiple tasks. These challenges preclude their applicability, and motivate our proposal of CATCH, a novel Context-bAsed meTa reinforcement learning (RL) algorithm for transferrable arChitecture searcH. The combination of meta-learning and RL allows CATCH to efficiently adapt to new tasks while being agnostic to search spaces. CATCH utilizes a probabilistic encoder to encode task properties into latent context variables, which then guide CATCH's controller to quickly "catch" top-performing networks. The contexts also assist a network evaluator in filtering inferior candidates and speed up learning. Extensive experiments demonstrate CATCH's universality and search efficiency over many other widely-recognized algorithms. It is also capable of handling cross-domain architecture search as competitive networks on ImageNet, COCO, and Cityscapes are identified. This is the first work to our knowledge that proposes an efficient transferrable NAS solution while maintaining robustness across various settings.

Results

TaskDatasetMetricValueModel
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Accuracy (Val)46.07CATCH-meta
Neural Architecture SearchNAS-Bench-201, ImageNet-16-120Search time (s)18000CATCH-meta
Neural Architecture SearchNAS-Bench-201, CIFAR-10Accuracy (Val)91.33CATCH-meta
Neural Architecture SearchNAS-Bench-201, CIFAR-100Accuracy (Val)72.57CATCH-meta
AutoMLNAS-Bench-201, ImageNet-16-120Accuracy (Val)46.07CATCH-meta
AutoMLNAS-Bench-201, ImageNet-16-120Search time (s)18000CATCH-meta
AutoMLNAS-Bench-201, CIFAR-10Accuracy (Val)91.33CATCH-meta
AutoMLNAS-Bench-201, CIFAR-100Accuracy (Val)72.57CATCH-meta

Related Papers

CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning2025-07-18VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Aligning Humans and Robots via Reinforcement Learning from Implicit Human Feedback2025-07-17VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Autonomous Resource Management in Microservice Systems via Reinforcement Learning2025-07-17