Weight-Entanglement Meets Gradient-Based Neural Architecture Search

Published: 30 Apr 2024, Last Modified: 05 Sept 2024AutoML 2024EveryoneRevisionsBibTeXCC BY 4.0
Keywords: Neural Architecture Search, Gradient-based, Efficiency
TL;DR: Studying weight entanglement for efficient gradient based neural architecture search
Abstract: Weight sharing is a fundamental concept in neural architecture search (NAS), enabling gradient-based methods to explore cell-based architectural spaces significantly faster than traditional blackbox approaches. In parallel, weight entanglement has emerged as a technique for more intricate parameter sharing amongst macro-architectural spaces. Since weight-entanglement is not directly compatible with gradient-based NAS methods, these two paradigms have largely developed independently in parallel sub-communities. This paper aims to bridge the gap between these sub-communities by proposing a novel scheme to adapt gradient-based methods for weight-entangled spaces. This enables us to conduct an in-depth comparative assessment and analysis of the performance of gradient-based NAS in weight-entangled search spaces. Our findings reveal that this integration of weight-entanglement and gradient-based NAS brings forth the various benefits of gradient-based methods, while preserving the memory efficiency of weight-entangled spaces. The code for our work is openly accessible at https://anon-github.automl.cc/r/TangleNAS-5BA5.
Submission Checklist: Yes
Broader Impact Statement: Yes
Paper Availability And License: Yes
Code Of Conduct: Yes
Optional Meta-Data For Green-AutoML: This blue field is just for structuring purposes and cannot be filled.
Steps For Environmental Footprint Reduction During Development: Our method by its nature is gradient based and very efficient (does not involve multiple trainings of a network)
CPU Hours: 80000
GPU Hours: 10000
Evaluation Metrics: Yes
Submission Number: 2
Loading