State Space Representations of Deep Neural Networks

06/11/2018
by   Michael Hauser, et al.
0

This paper deals with neural networks as dynamical systems governed by differential or difference equations. It shows that the introduction of skip connections into network architectures, such as residual networks and dense networks, turns a system of static equations into a system of dynamical equations with varying levels of smoothness on the layer-wise transformations. Closed form solutions for the state space representations of general dense networks, as well as k^th order smooth networks, are found in general settings. Furthermore, it is shown that imposing k^th order smoothness on a network architecture with d-many nodes per layer increases the state space dimension by a multiple of k, and so the effective embedding dimension of the data manifold is k · d-many dimensions. It follows that network architectures of these types reduce the number of parameters needed to maintain the same embedding dimension by a factor of k^2 when compared to an equivalent first-order, residual network, significantly motivating the development of network architectures of these types. Numerical simulations were run to validate parts of the developed theory.

READ FULL TEXT
research
10/21/2019

Variational Integrator Networks for Physically Meaningful Embeddings

Learning workable representations of dynamical systems is becoming an in...
research
02/20/2020

Comparing recurrent and convolutional neural networks for predicting wave propagation

Dynamical systems can be modelled by partial differential equations and ...
research
08/02/2023

Embedding Capabilities of Neural ODEs

A class of neural networks that gained particular interest in the last y...
research
08/11/2021

Learning strange attractors with reservoir systems

This paper shows that the celebrated Embedding Theorem of Takens is a pa...
research
09/27/2018

Smooth Inter-layer Propagation of Stabilized Neural Networks for Classification

Recent work has studied the reasons for the remarkable performance of de...
research
03/29/2021

Translating Numerical Concepts for PDEs into Neural Architectures

We investigate what can be learned from translating numerical algorithms...
research
02/12/2019

Capacity allocation analysis of neural networks: A tool for principled architecture design

Designing neural network architectures is a task that lies somewhere bet...

Please sign up or login with your details

Forgot password? Click here to reset