ExplainFix: Explainable Spatially Fixed Deep Networks

03/18/2023
by   Alex Gaudio, et al.
0

Is there an initialization for deep networks that requires no learning? ExplainFix adopts two design principles: the "fixed filters" principle that all spatial filter weights of convolutional neural networks can be fixed at initialization and never learned, and the "nimbleness" principle that only few network parameters suffice. We contribute (a) visual model-based explanations, (b) speed and accuracy gains, and (c) novel tools for deep convolutional neural networks. ExplainFix gives key insights that spatially fixed networks should have a steered initialization, that spatial convolution layers tend to prioritize low frequencies, and that most network parameters are not necessary in spatially fixed models. ExplainFix models have up to 100x fewer spatial filter kernels than fully learned models and matching or improved accuracy. Our extensive empirical analysis confirms that ExplainFix guarantees nimbler models (train up to 17% faster with channel pruning), matching or improved predictive performance (spanning 13 distinct baseline models, four architectures and two medical image datasets), improved robustness to larger learning rate, and robustness to varying model size. We are first to demonstrate that all spatial filters in state-of-the-art convolutional deep networks can be fixed at initialization, not learned.

READ FULL TEXT

page 4

page 7

page 12

page 21

page 24

page 25

page 26

page 28

research
11/30/2017

Spatially-Adaptive Filter Units for Deep Neural Networks

Classical deep convolutional networks increase receptive field size by e...
research
07/25/2022

Trainability Preserving Neural Structured Pruning

Several recent works empirically find finetuning learning rate is critic...
research
07/25/2019

Filter Bank Regularization of Convolutional Neural Networks

Regularization techniques are widely used to improve the generality, rob...
research
10/07/2022

Understanding the Covariance Structure of Convolutional Filters

Neural network weights are typically initialized at random from univaria...
research
02/18/2016

RandomOut: Using a convolutional gradient norm to rescue convolutional filters

Filters in convolutional neural networks are sensitive to their initiali...
research
06/14/2018

Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks

In recent years, state-of-the-art methods in computer vision have utiliz...
research
08/17/2022

Performance Evaluation of Selective Fixed-filter Active Noise Control based on Different Convolutional Neural Networks

Due to its rapid response time and a high degree of robustness, the sele...

Please sign up or login with your details

Forgot password? Click here to reset