To Share or Not To Share: A Comprehensive Appraisal of Weight-Sharing
Aloïs Pourchot, Alexis Ducarouge, Olivier Sigaud
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/apourchot/to_share_or_not_to_shareOfficialpytorch★ 7
Abstract
Weight-sharing (WS) has recently emerged as a paradigm to accelerate the automated search for efficient neural architectures, a process dubbed Neural Architecture Search (NAS). Although very appealing, this framework is not without drawbacks and several works have started to question its capabilities on small hand-crafted benchmarks. In this paper, we take advantage of the dataset to challenge the efficiency of WS on a representative search space. By comparing a SOTA WS approach to a plain random search we show that, despite decent correlations between evaluations using weight-sharing and standalone ones, WS is only rarely significantly helpful to NAS. In particular we highlight the impact of the search space itself on the benefits.