Architopes: An Architecture Modification for Composite Pattern Learning, Increased Expressiveness, and Reduced Training Time

06/24/2020
by   Anastasis Kratsios, et al.
0

We introduce a simple neural network architecture modification that enables composite pattern learning, increases expressiveness, and reduces training time. This expressibility improvement is explained by the density of the modified architecture in a new refined local L^p-space describing composite patterns. In contrast, most feed-forward neural network architectures with sigmoid activation functions are shown not to be dense in this space. In practice, restrictions have to be placed on the dimension of any architecture's parameter space. L^1 approximation bounds are obtained in terms of the number of the trainable parameters. Likewise, convergence guarantees are obtained as the imposed restrictions are asymptotically removed. By exploiting the new architecture's structure, a parallelizable training meta-algorithm is provided, and numerical evaluations are made using the California housing dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/29/2020

Overcoming The Limitations of Neural Networks in Composite-Pattern Learning with Architopes

The effectiveness of neural networks in solving complex problems is well...
research
07/03/2019

Neural Network Architecture Search with Differentiable Cartesian Genetic Programming for Regression

The ability to design complex neural network architectures which enable ...
research
10/10/2018

Automatic Configuration of Deep Neural Networks with EGO

Designing the architecture for an artificial neural network is a cumbers...
research
03/01/2023

Multi-task neural networks by learned contextual inputs

This paper explores learned-context neural networks. It is a multi-task ...
research
02/28/2011

Improving the character recognition efficiency of feed forward BP neural network

This work is focused on improving the character recognition capability o...
research
01/04/2022

An unfeasability view of neural network learning

We define the notion of a continuously differentiable perfect learning a...
research
04/14/2023

The R-mAtrIx Net

We provide a novel Neural Network architecture that can: i) output R-mat...

Please sign up or login with your details

Forgot password? Click here to reset