Towards NNGP-guided Neural Architecture Search

11/11/2020
by   Daniel S. Park, et al.
6

The predictions of wide Bayesian neural networks are described by a Gaussian process, known as the Neural Network Gaussian Process (NNGP). Analytic forms for NNGP kernels are known for many models, but computing the exact kernel for convolutional architectures is prohibitively expensive. One can obtain effective approximations of these kernels through Monte-Carlo estimation using finite networks at initialization. Monte-Carlo NNGP inference is orders-of-magnitude cheaper in FLOPs compared to gradient descent training when the dataset size is small. Since NNGP inference provides a cheap measure of performance of a network architecture, we investigate its potential as a signal for neural architecture search (NAS). We compute the NNGP performance of approximately 423k networks in the NAS-bench 101 dataset on CIFAR-10 and compare its utility against conventional performance measures obtained by shortened gradient-based training. We carry out a similar analysis on 10k randomly sampled networks in the mobile neural architecture search (MNAS) space for ImageNet. We discover comparative advantages of NNGP-based metrics, and discuss potential applications. In particular, we propose that NNGP performance is an inexpensive signal independent of metrics obtained from training that can either be used for reducing big search spaces, or improving training-based performance measures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/25/2019

BANANAS: Bayesian Optimization with Neural Architectures for Neural Architecture Search

Neural Architecture Search (NAS) has seen an explosion of research in th...
research
10/12/2018

Graph HyperNetworks for Neural Architecture Search

Neural architecture search (NAS) automatically finds the best task-speci...
research
12/16/2022

From Xception to NEXcepTion: New Design Decisions and Neural Architecture Search

In this paper, we present a modified Xception architecture, the NEXcepTi...
research
06/13/2020

Optimal Transport Kernels for Sequential and Parallel Neural Architecture Search

Neural architecture search (NAS) automates the design of deep neural net...
research
10/16/2021

GradSign: Model Performance Inference with Theoretical Insights

A key challenge in neural architecture search (NAS) is quickly inferring...
research
03/17/2021

The Untapped Potential of Off-the-Shelf Convolutional Neural Networks

Over recent years, a myriad of novel convolutional network architectures...
research
11/26/2021

KNAS: Green Neural Architecture Search

Many existing neural architecture search (NAS) solutions rely on downstr...

Please sign up or login with your details

Forgot password? Click here to reset