Teacher Guided Architecture Search

08/04/2018
by   Pouya Bashivan, et al.
0

Strong improvements in network performance in vision tasks have resulted from the search of alternative network architectures, and prior work has shown that this search process can be automated and guided by evaluating candidate network performance following limited training (Performance Guided Architecture Search or PGAS). However, because of the large architecture search spaces and the high computational cost associated with evaluating each candidate model, further gains in computational efficiency are needed. Here we present a method termed Teacher Guided Search for Architectures by Generation and Evaluation (TG-SAGE) that produces up to an order of magnitude in search efficiency over PGAS methods. Specifically, TG-SAGE guides each step of the architecture search by evaluating the similarity of internal representations of the candidate networks with those of the (fixed) teacher network. We show that this procedure leads to significant reduction in required per-sample training and that, this advantage holds for two different search spaces of architectures, and two different search algorithms. We further show that in the space of convolutional cells for visual categorization, TG-SAGE finds a cell structure with similar performance as was previously found using other methods but at a total computational cost that is two orders of magnitude lower than Neural Architecture Search (NAS) and more than four times lower than progressive neural architecture search (PNAS). These results suggest that TG-SAGE can be used to accelerate network architecture search in cases where one has access to some or all of the internal representations of a teacher network of interest, such as the brain.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2019

Probabilistic Neural Architecture Search

In neural architecture search (NAS), the space of neural network archite...
research
11/29/2019

Blockwisely Supervised Neural Architecture Search with Knowledge Distillation

Neural Architecture Search (NAS), aiming at automatically designing netw...
research
02/01/2019

Learnable Embedding Space for Efficient Neural Architecture Compression

We propose a method to incrementally learn an embedding space over the d...
research
01/17/2020

Up to two billion times acceleration of scientific simulations with deep neural architecture search

Computer simulations are invaluable tools for scientific discovery. Howe...
research
08/19/2019

Architecture Search by Estimation of Network Structure Distributions

The influence of deep learning is continuously expanding across differen...
research
05/31/2021

Memory-Efficient Differentiable Transformer Architecture Search

Differentiable architecture search (DARTS) is successfully applied in ma...
research
12/30/2019

Neural Architecture Search on Acoustic Scene Classification

Convolutional neural networks are widely adopted in Acoustic Scene Class...

Please sign up or login with your details

Forgot password? Click here to reset