Perceptron Theory for Predicting the Accuracy of Neural Networks

12/14/2020
by   Denis Kleyko, et al.
0

Many neural network models have been successful at classification problems, but their operation is still treated as a black box. Here, we developed a theory for one-layer perceptrons that can predict performance on classification tasks. This theory is a generalization of an existing theory for predicting the performance of Echo State Networks and connectionist models for symbolic reasoning known as Vector Symbolic Architectures. In this paper, we first show that the proposed perceptron theory can predict the performance of Echo State Networks, which could not be described by the previous theory. Second, we apply our perceptron theory to the last layers of shallow randomly connected and deep multi-layer networks. The full theory is based on Gaussian statistics, but it is analytically intractable. We explore numerical methods to predict network performance for problems with a small number of classes. For problems with a large number of classes, we investigate stochastic sampling methods and a tractable approximation to the full theory. The quality of predictions is assessed in three experimental settings, using reservoir computing networks on a memorization task, shallow randomly connected networks on a collection of classification datasets, and deep convolutional networks with the ImageNet dataset. This study offers a simple, bipartite approach to understand deep neural networks: the input is encoded by the last-but-one layers into a high-dimensional representation. This representation is mapped through the weights of the last layer into the postsynaptic sums of the output neurons. Specifically, the proposed perceptron theory uses the mean vector and covariance matrix of the postsynaptic sums to compute classification accuracies for the different classes. The first two moments of the distribution of the postsynaptic sums can predict the overall network performance quite accurately.

READ FULL TEXT
research
02/15/2019

Efficient Deep Learning of GMMs

We show that a collection of Gaussian mixture models (GMMs) in R^n can b...
research
11/09/2018

A Convergence Theory for Deep Learning via Over-Parameterization

Deep neural networks (DNNs) have demonstrated dominating performance in ...
research
12/02/2019

Capacity of the covariance perceptron

The classical perceptron is a simple neural network that performs a bina...
research
07/08/2022

Braid-based architecture search

In this article, we propose the approach to structural optimization of n...
research
07/02/2022

Parameter efficient dendritic-tree neurons outperform perceptrons

Biological neurons are more powerful than artificial perceptrons, in par...
research
02/04/2022

A note on the complex and bicomplex valued neural networks

In this paper we first write a proof of the perceptron convergence algor...
research
08/29/2022

Rosenblatt's first theorem and frugality of deep learning

First Rosenblatt's theorem about omnipotence of shallow networks states ...

Please sign up or login with your details

Forgot password? Click here to reset