Search Spaces for Neural Model Training

05/27/2021
by   Darko Stosic, et al.
0

While larger neural models are pushing the boundaries of what deep learning can do, often more weights are needed to train models rather than to run inference for tasks. This paper seeks to understand this behavior using search spaces – adding weights creates extra degrees of freedom that form new paths for optimization (or wider search spaces) rendering neural model training more effective. We then show how we can augment search spaces to train sparse models attaining competitive scores across dozens of deep learning workloads. They are also are tolerant of structures targeting current hardware, opening avenues for training and inference acceleration. Our work encourages research to explore beyond massive neural models being used today.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2011

The Ariadne's Clew Algorithm

We present a new approach to path planning, called the "Ariadne's clew a...
research
07/02/2023

Tools for Verifying Neural Models' Training Data

It is important that consumers and regulators can verify the provenance ...
research
09/25/2021

Profiling Neural Blocks and Design Spaces for Mobile Neural Architecture Search

Neural architecture search automates neural network design and has achie...
research
11/21/2022

Exploring Physical Latent Spaces for Deep Learning

We explore training deep neural network models in conjunction with physi...
research
08/17/2021

Learning C to x86 Translation: An Experiment in Neural Compilation

Deep learning has had a significant impact on many fields. Recently, cod...
research
07/27/2017

Effective Inference for Generative Neural Parsing

Generative neural models have recently achieved state-of-the-art results...
research
08/08/2020

A novel residual whitening based training to avoid overfitting

In this paper we demonstrate that training models to minimize the autoco...

Please sign up or login with your details

Forgot password? Click here to reset