Dive into Layers: Neural Network Capacity Bounding using Algebraic Geometry

09/03/2021
by   Ji Yang, et al.
0

The empirical results suggest that the learnability of a neural network is directly related to its size. To mathematically prove this, we borrow a tool in topological algebra: Betti numbers to measure the topological geometric complexity of input data and the neural network. By characterizing the expressive capacity of a neural network with its topological complexity, we conduct a thorough analysis and show that the network's expressive capacity is limited by the scale of its layers. Further, we derive the upper bounds of the Betti numbers on each layer within the network. As a result, the problem of architecture selection of a neural network is transformed to determining the scale of the network that can represent the input data complexity. With the presented results, the architecture selection of a fully connected network boils down to choosing a suitable size of the network such that it equips the Betti numbers that are not smaller than the Betti numbers of the input data. We perform the experiments on a real-world dataset MNIST and the results verify our analysis and conclusion. The code is publicly available.

READ FULL TEXT
research
02/13/2018

On Characterizing the Capacity of Neural Networks using Algebraic Topology

The learnability of different neural architectures can be characterized ...
research
12/28/2019

Classifying topological sector via machine learning

We employ a machine learning technique for an estimate of the topologica...
research
05/25/2023

Data Topology-Dependent Upper Bounds of Neural Network Widths

This paper investigates the relationship between the universal approxima...
research
11/24/2015

Dynamic Capacity Networks

We introduce the Dynamic Capacity Network (DCN), a neural network that c...
research
07/17/2017

The PSLQ Algorithm for Empirical Data

The celebrated integer relation finding algorithm PSLQ has been successf...
research
12/14/2016

Deep Function Machines: Generalized Neural Networks for Topological Layer Expression

In this paper we propose a generalization of deep neural networks called...
research
01/03/2021

Algorithmic Complexities in Backpropagation and Tropical Neural Networks

In this note, we propose a novel technique to reduce the algorithmic com...

Please sign up or login with your details

Forgot password? Click here to reset