Neural Functional Transformers

05/22/2023
by   Allan Zhou, et al.
0

The recent success of neural networks as implicit representation of data has driven growing interest in neural functionals: models that can process other neural networks as input by operating directly over their weight spaces. Nevertheless, constructing expressive and efficient neural functional architectures that can handle high-dimensional weight-space objects remains challenging. This paper uses the attention mechanism to define a novel set of permutation equivariant weight-space layers and composes them into deep equivariant models called neural functional Transformers (NFTs). NFTs respect weight-space permutation symmetries while incorporating the advantages of attention, which have exhibited remarkable success across multiple domains. In experiments processing the weights of feedforward MLPs and CNNs, we find that NFTs match or exceed the performance of prior weight-space methods. We also leverage NFTs to develop Inr2Array, a novel method for computing permutation invariant latent representations from the weights of implicit neural representations (INRs). Our proposed method improves INR classification accuracy by up to +17% over existing methods. We provide an implementation of our layers at https://github.com/AllanYangZhou/nfn.

READ FULL TEXT

page 7

page 15

research
02/27/2023

Permutation Equivariant Neural Functionals

This work studies the design of neural networks that can process the wei...
research
10/27/2021

Transformers Generalize DeepSets and Can be Extended to Graphs and Hypergraphs

We present a generalization of Transformers to any-order permutation inv...
research
01/30/2023

Equivariant Architectures for Learning in Deep Weight Spaces

Designing machine learning architectures for processing neural networks ...
research
09/18/2023

Latent assimilation with implicit neural representations for unknown dynamics

Data assimilation is crucial in a wide range of applications, but it oft...
research
10/08/2021

Pathologies in priors and inference for Bayesian transformers

In recent years, the transformer has established itself as a workhorse i...
research
11/23/2022

Generalizable Implicit Neural Representations via Instance Pattern Composers

Despite recent advances in implicit neural representations (INRs), it re...
research
10/05/2020

Are Neural Nets Modular? Inspecting Functional Modularity Through Differentiable Weight Masks

Neural networks (NNs) whose subnetworks implement reusable functions are...

Please sign up or login with your details

Forgot password? Click here to reset