Hyperplane Arrangements of Trained ConvNets Are Biased

03/17/2020
by   Matteo Gamba, et al.
10

We investigate the geometric properties of the functions learned by trained ConvNets in the preactivation space of their convolutional layers, by performing an empirical study of hyperplane arrangements induced by a convolutional layer. We introduce statistics over the weights of a trained network to study local arrangements and relate them to the training dynamics. We observe that trained ConvNets show a significant statistical bias towards regular hyperplane configurations. Furthermore, we find that layers showing biased configurations are critical to validation performance for the architectures considered, trained on CIFAR10, CIFAR100 and ImageNet.

READ FULL TEXT

page 2

page 14

page 16

page 17

page 20

research
06/16/2020

Reusing Trained Layers of Convolutional Neural Networks to Shorten Hyperparameters Tuning Time

Hyperparameters tuning is a time-consuming approach, particularly when t...
research
05/26/2018

The Singular Values of Convolutional Layers

We characterize the singular values of the linear transformation associa...
research
02/28/2017

ShaResNet: reducing residual network parameter number by sharing weights

Deep Residual Networks have reached the state of the art in many image p...
research
11/19/2018

Biologically plausible deep learning

Building on the model proposed in Lillicrap et. al. we show that deep ne...
research
03/23/2016

A guide to convolution arithmetic for deep learning

We introduce a guide to help deep learning practitioners understand and ...
research
06/10/2021

Transformed CNNs: recasting pre-trained convolutional layers with self-attention

Vision Transformers (ViT) have recently emerged as a powerful alternativ...
research
06/19/2023

Algorithms of Sampling-Frequency-Independent Layers for Non-integer Strides

In this paper, we propose algorithms for handling non-integer strides in...

Please sign up or login with your details

Forgot password? Click here to reset