AttentiveNAS: Improving Neural Architecture Search via Attentive Sampling

11/18/2020
by   Dilin Wang, et al.
0

Neural architecture search (NAS) has shown great promise designing state-of-the-art (SOTA) models that are both accurate and fast. Recently, two-stage NAS, e.g. BigNAS, decouples the model training and searching process and achieves good search efficiency. Two-stage NAS requires sampling from the search space during training, which directly impacts the accuracy of the final searched models. While uniform sampling has been widely used for simplicity, it is agnostic of the model performance Pareto front, which are the main focus in the search process, and thus, misses opportunities to further improve the model accuracy. In this work, we propose AttentiveNAS that focuses on sampling the networks to improve the performance Pareto. We also propose algorithms to efficiently and effectively identify the networks on the Pareto during training. Without extra re-training or post-processing, we can simultaneously obtain a large number of networks across a wide range of FLOPs. Our discovered model family, AttentiveNAS models, achieves top-1 accuracy from 77.3 on ImageNet, and outperforms SOTA models, including BigNAS, Once-for-All networks and FBNetV3. We also achieve ImageNet accuracy of 80.1 MFLOPs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/06/2021

Going Beyond Neural Architecture Search with Sampling-based Neural Ensemble Search

Recently, Neural Architecture Search (NAS) has been widely applied to au...
research
03/24/2020

BigNAS: Scaling Up Neural Architecture Search with Big Single-Stage Models

Neural architecture search (NAS) has shown promising results discovering...
research
03/17/2023

ElasticViT: Conflict-aware Supernet Training for Deploying Fast Vision Transformer on Diverse Mobile Devices

Neural Architecture Search (NAS) has shown promising performance in the ...
research
02/16/2021

AlphaNet: Improved Training of Supernet with Alpha-Divergence

Weight-sharing neural architecture search (NAS) is an effective techniqu...
research
08/23/2022

FocusFormer: Focusing on What We Need via Architecture Sampler

Vision Transformers (ViTs) have underpinned the recent breakthroughs in ...
research
11/22/2020

FP-NAS: Fast Probabilistic Neural Architecture Search

Differential Neural Architecture Search (NAS) requires all layer choices...
research
07/17/2023

ShiftNAS: Improving One-shot NAS via Probability Shift

One-shot Neural architecture search (One-shot NAS) has been proposed as ...

Please sign up or login with your details

Forgot password? Click here to reset