Dataless Model Selection with the Deep Frame Potential

03/30/2020
by   Calvin Murdock, et al.
0

Choosing a deep neural network architecture is a fundamental problem in applications that require balancing performance and parameter efficiency. Standard approaches rely on ad-hoc engineering or computationally expensive validation on a specific dataset. We instead attempt to quantify networks by their intrinsic capacity for unique and robust representations, enabling efficient architecture comparisons without requiring any data. Building upon theoretical connections between deep learning and sparse approximation, we propose the deep frame potential: a measure of coherence that is approximately related to representation stability but has minimizers that depend only on network structure. This provides a framework for jointly quantifying the contributions of architectural hyper-parameters such as depth, width, and skip connections. We validate its use as a criterion for model selection and demonstrate correlation with generalization error on a variety of common residual and densely connected network architectures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/10/2021

Reframing Neural Networks: Deep Structure in Overcomplete Representations

In comparison to classical shallow representation learning techniques, d...
research
02/01/2023

Width and Depth Limits Commute in Residual Networks

We show that taking the width and depth to infinity in a deep neural net...
research
05/15/2019

Automatic Model Selection for Neural Networks

Neural networks and deep learning are changing the way that artificial i...
research
02/11/2018

Neural Architecture Search with Bayesian Optimisation and Optimal Transport

Bayesian Optimisation (BO) refers to a class of methods for global optim...
research
07/02/2021

Neural Network Layer Algebra: A Framework to Measure Capacity and Compression in Deep Learning

We present a new framework to measure the intrinsic properties of (deep)...
research
10/06/2021

SIRe-Networks: Skip Connections over Interlaced Multi-Task Learning and Residual Connections for Structure Preserving Object Classification

Improving existing neural network architectures can involve several desi...
research
07/05/2022

A Densely Interconnected Network for Deep Learning Accelerated MRI

Objective: To improve accelerated MRI reconstruction through a densely c...

Please sign up or login with your details

Forgot password? Click here to reset