Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis

05/11/2022
by   Wuyang Chen, et al.
12

Advanced deep neural networks (DNNs), designed by either human or AutoML algorithms, are growing increasingly complex. Diverse operations are connected by complicated connectivity patterns, e.g., various types of skip connections. Those topological compositions are empirically effective and observed to smooth the loss landscape and facilitate the gradient flow in general. However, it remains elusive to derive any principled understanding of their effects on the DNN capacity or trainability, and to understand why or in which aspect one specific connectivity pattern is better than another. In this work, we theoretically characterize the impact of connectivity patterns on the convergence of DNNs under gradient descent training in fine granularity. By analyzing a wide network's Neural Network Gaussian Process (NNGP), we are able to depict how the spectrum of an NNGP kernel propagates through a particular connectivity pattern, and how that affects the bound of convergence rates. As one practical implication of our results, we show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate, and significantly accelerate the large-scale neural architecture search without any overhead. Codes will be released at https://github.com/chenwydj/architecture_convergence.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2018

Deep Neural Networks Learn Non-Smooth Functions Effectively

We theoretically discuss why deep neural networks (DNNs) performs better...
research
04/03/2020

Gradient Centralization: A New Optimization Technique for Deep Neural Networks

Optimization techniques are of great importance to effectively and effic...
research
09/14/2022

Optimal Connectivity through Network Gradients for the Restricted Boltzmann Machine

Leveraging sparse networks to connect successive layers in deep neural n...
research
02/12/2021

Neural Architecture Search as Program Transformation Exploration

Improving the performance of deep neural networks (DNNs) is important to...
research
06/07/2022

Integrating Random Effects in Deep Neural Networks

Modern approaches to supervised learning like deep neural networks (DNNs...
research
11/26/2021

KNAS: Green Neural Architecture Search

Many existing neural architecture search (NAS) solutions rely on downstr...
research
07/28/2018

MaskConnect: Connectivity Learning by Gradient Descent

Although deep networks have recently emerged as the model of choice for ...

Please sign up or login with your details

Forgot password? Click here to reset