Accelerating Evolutionary Neural Architecture Search via Multi-Fidelity Evaluation

08/10/2021
by   Shangshang Yang, et al.
4

Evolutionary neural architecture search (ENAS) has recently received increasing attention by effectively finding high-quality neural architectures, which however consumes high computational cost by training the architecture encoded by each individual for complete epochs in individual evaluation. Numerous ENAS approaches have been developed to reduce the evaluation cost, but it is often difficult for most of these approaches to achieve high evaluation accuracy. To address this issue, in this paper we propose an accelerated ENAS via multifidelity evaluation termed MFENAS, where the individual evaluation cost is significantly reduced by training the architecture encoded by each individual for only a small number of epochs. The balance between evaluation cost and evaluation accuracy is well maintained by suggesting a multi-fidelity evaluation, which identifies the potentially good individuals that cannot survive from previous generations by integrating multiple evaluations under different numbers of training epochs. For high diversity of neural architectures, a population initialization strategy is devised to produce different neural architectures varying from ResNet-like architectures to Inception-like ones. Experimental results on CIFAR-10 show that the architecture obtained by the proposed MFENAS achieves a 2.39 at the cost of only 0.6 GPU days on one NVIDIA 2080TI GPU, demonstrating the superiority of the proposed MFENAS over state-of-the-art NAS approaches in terms of both computational cost and architecture quality. The architecture obtained by the proposed MFENAS is then transferred to CIFAR-100 and ImageNet, which also exhibits competitive performance to the architectures obtained by existing NAS approaches. The source code of the proposed MFENAS is available at https://github.com/DevilYangS/MFENAS/.

READ FULL TEXT

page 2

page 4

page 9

page 11

page 12

page 13

page 14

page 15

research
02/24/2020

Semi-Supervised Neural Architecture Search

Neural architecture search (NAS) relies on a good controller to generate...
research
06/09/2021

Accelerating Neural Architecture Search via Proxy Data

Despite the increasing interest in neural architecture search (NAS), the...
research
08/16/2023

ResBuilder: Automated Learning of Depth with Residual Structures

In this work, we develop a neural architecture search algorithm, termed ...
research
01/17/2018

FastNet

Inception and the Resnet family of Convolutional Neural Network archi-te...
research
08/09/2021

BenchENAS: A Benchmarking Platform for Evolutionary Neural Architecture Search

Neural architecture search (NAS), which automatically designs the archit...
research
03/23/2022

Training-free Transformer Architecture Search

Recently, Vision Transformer (ViT) has achieved remarkable success in se...
research
07/26/2023

YOLOBench: Benchmarking Efficient Object Detectors on Embedded Systems

We present YOLOBench, a benchmark comprised of 550+ YOLO-based object de...

Please sign up or login with your details

Forgot password? Click here to reset