Embeddings between Barron spaces with higher order activation functions

05/25/2023
by   Tjeerd Jan Heeringa, et al.
0

The approximation properties of infinitely wide shallow neural networks heavily depend on the choice of the activation function. To understand this influence, we study embeddings between Barron spaces with different activation functions. These embeddings are proven by providing push-forward maps on the measures μ used to represent functions f. An activation function of particular interest is the rectified power unit (RePU) given by RePU_s(x)=max(0,x)^s. For many commonly used activation functions, the well-known Taylor remainder theorem can be used to construct a push-forward map, which allows us to prove the embedding of the associated Barron space into a Barron space with a RePU as activation function. Moreover, the Barron spaces associated with the RePU_s have a hierarchical structure similar to the Sobolev spaces H^m.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2023

STL: A Signed and Truncated Logarithm Activation Function for Neural Networks

Activation functions play an essential role in neural networks. They pro...
research
08/02/2018

The Quest for the Golden Activation Function

Deep Neural Networks have been shown to be beneficial for a variety of t...
research
01/27/2021

Kähler Geometry of Quiver Varieties and Machine Learning

We develop an algebro-geometric formulation for neural networks in machi...
research
06/29/2023

Why Shallow Networks Struggle with Approximating and Learning High Frequency: A Numerical Study

In this work, a comprehensive numerical study involving analysis and exp...
research
01/30/2020

Conditioning Autoencoder Latent Spaces for Real-Time Timbre Interpolation and Synthesis

We compare standard autoencoder topologies' performances for timbre gene...
research
06/09/2020

A Note on Deepfake Detection with Low-Resources

Deepfakes are videos that include changes, quite often substituting face...
research
10/15/2020

Review and Comparison of Commonly Used Activation Functions for Deep Neural Networks

The primary neural networks decision-making units are activation functio...

Please sign up or login with your details

Forgot password? Click here to reset