Globally injective and bijective neural operators

06/06/2023
by   Takashi Furuya, et al.
0

Recently there has been great interest in operator learning, where networks learn operators between function spaces from an essentially infinite-dimensional perspective. In this work we present results for when the operators learned by these networks are injective and surjective. As a warmup, we combine prior work in both the finite-dimensional ReLU and operator learning setting by giving sharp conditions under which ReLU layers with linear neural operators are injective. We then consider the case the case when the activation function is pointwise bijective and obtain sufficient conditions for the layer to be injective. We remark that this question, while trivial in the finite-rank case, is subtler in the infinite-rank case and is proved using tools from Fredholm theory. Next, we prove that our supplied injective neural operators are universal approximators and that their implementation, with finite-rank neural networks, are still injective. This ensures that injectivity is not `lost' in the transcription from analytical operators to their finite-rank implementation with networks. Finally, we conclude with an increase in abstraction and consider general conditions when subnetworks, which may be many layers deep, are injective and surjective and provide an exact inversion from a `linearization.' This section uses general arguments from Fredholm theory and Leray-Schauder degree theory for non-linear integral equations to analyze the mapping properties of neural operators in function spaces. These results apply to subnetworks formed from the layers considered in this work, under natural conditions. We believe that our work has applications in Bayesian UQ where injectivity enables likelihood estimation and in inverse problems where surjectivity and injectivity corresponds to existence and uniqueness, respectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/19/2021

Neural Operator: Learning Maps Between Function Spaces

The classical development of neural networks has primarily focused on le...
research
11/21/2021

A Pseudo-Inverse for Nonlinear Operators

The Moore-Penrose inverse is widely used in physics, statistics and vari...
research
06/15/2022

Local Identifiability of Deep ReLU Neural Networks: the Theory

Is a sample rich enough to determine, at least locally, the parameters o...
research
05/05/2021

Two-layer neural networks with values in a Banach space

We study two-layer neural networks whose domain and range are Banach spa...
research
01/29/2019

Learning Schatten--Von Neumann Operators

We study the learnability of a class of compact operators known as Schat...
research
07/06/2022

On Non-Linear operators for Geometric Deep Learning

This work studies operators mapping vector and scalar fields defined ove...
research
10/15/2018

A unified approach to calculation of information operators in semiparametric models

The infinite-dimensional information operator for the nuisance parameter...

Please sign up or login with your details

Forgot password? Click here to reset