On Privileged and Convergent Bases in Neural Network Representations

07/24/2023
by   Davis Brown, et al.
0

In this study, we investigate whether the representations learned by neural networks possess a privileged and convergent basis. Specifically, we examine the significance of feature directions represented by individual neurons. First, we establish that arbitrary rotations of neural representations cannot be inverted (unlike linear networks), indicating that they do not exhibit complete rotational invariance. Subsequently, we explore the possibility of multiple bases achieving identical performance. To do this, we compare the bases of networks trained with the same parameters but with varying random initializations. Our study reveals two findings: (1) Even in wide networks such as WideResNets, neural networks do not converge to a unique basis; (2) Basis correlation increases significantly when a few early layers of the network are frozen identically. Furthermore, we analyze Linear Mode Connectivity, which has been studied as a measure of basis correlation. Our findings give evidence that while Linear Mode Connectivity improves with increased network width, this improvement is not due to an increase in basis correlation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/14/2018

Insights on representational similarity in neural networks with canonical correlation

Comparing different neural network representations and determining how r...
research
11/29/2022

Infinite-width limit of deep linear neural networks

This paper studies the infinite-width limit of deep linear neural networ...
research
11/24/2015

Convergent Learning: Do different neural networks learn the same representations?

Recent success in training deep neural networks have prompted active inv...
research
03/09/2021

Discrete Function Bases and Convolutional Neural Networks

We discuss the notion of "discrete function bases" with a particular foc...
research
06/14/2021

Revisiting Model Stitching to Compare Neural Representations

We revisit and extend model stitching (Lenc Vedaldi 2015) as a metho...
research
09/25/2019

Wider Networks Learn Better Features

Transferability of learned features between tasks can massively reduce t...
research
12/31/2018

Path-Invariant Map Networks

Optimizing a network of maps among a collection of objects/domains (or m...

Please sign up or login with your details

Forgot password? Click here to reset