SOTAVerified

SGAS: Sequential Greedy Architecture Search

2019-11-30CVPR 2020Code Available0· sign in to hype

Guohao Li, Guocheng Qian, Itzel C. Delgadillo, Matthias Müller, Ali Thabet, Bernard Ghanem

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Architecture design has become a crucial component of successful deep learning. Recent progress in automatic neural architecture search (NAS) shows a lot of promise. However, discovered architectures often fail to generalize in the final evaluation. Architectures with a higher validation accuracy during the search phase may perform worse in the evaluation. Aiming to alleviate this common issue, we introduce sequential greedy architecture search (SGAS), an efficient method for neural architecture search. By dividing the search procedure into sub-problems, SGAS chooses and prunes candidate operations in a greedy fashion. We apply SGAS to search architectures for Convolutional Neural Networks (CNN) and Graph Convolutional Networks (GCN). Extensive experiments show that SGAS is able to find state-of-the-art architectures for tasks such as image classification, point cloud classification and node classification in protein-protein interaction graphs with minimal computational cost. Please visit https://www.deepgcns.org/auto/sgas for more information about SGAS.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-10SGASTop-1 Error Rate2.39Unverified
ImageNetSGASTop-1 Error Rate24.1Unverified

Reproductions