Versatile Neural Processes for Learning Implicit Neural Representations

01/21/2023
by   Zongyu Guo, et al.
0

Representing a signal as a continuous function parameterized by neural network (a.k.a. Implicit Neural Representations, INRs) has attracted increasing attention in recent years. Neural Processes (NPs), which model the distributions over functions conditioned on partial observations (context set), provide a practical solution for fast inference of continuous functions. However, existing NP architectures suffer from inferior modeling capability for complex signals. In this paper, we propose an efficient NP framework dubbed Versatile Neural Processes (VNP), which largely increases the capability of approximating functions. Specifically, we introduce a bottleneck encoder that produces fewer and informative context tokens, relieving the high computational cost while providing high modeling capability. At the decoder side, we hierarchically learn multiple global latent variables that jointly model the global structure and the uncertainty of a function, enabling our model to capture the distribution of complex signals. We demonstrate the effectiveness of the proposed VNP on a variety of tasks involving 1D, 2D and 3D signals. Particularly, our method shows promise in learning accurate INRs w.r.t. a 3D scene without further finetuning.

READ FULL TEXT
research
06/17/2020

Implicit Neural Representations with Periodic Activation Functions

Implicitly defined, continuous, differentiable signal representations pa...
research
10/27/2021

Meta-Learning Sparse Implicit Neural Representations

Implicit neural representations are a promising new avenue of representi...
research
03/23/2023

Continuous Indeterminate Probability Neural Network

This paper introduces a general model called CIPNN - Continuous Indeterm...
research
03/31/2022

3D Equivariant Graph Implicit Functions

In recent years, neural implicit representations have made remarkable pr...
research
07/09/2021

InfoVAEGAN : learning joint interpretable representations by information maximization and maximum likelihood

Learning disentangled and interpretable representations is an important ...
research
11/15/2022

Latent Bottlenecked Attentive Neural Processes

Neural Processes (NPs) are popular methods in meta-learning that can est...
research
09/06/2023

ResFields: Residual Neural Fields for Spatiotemporal Signals

Neural fields, a category of neural networks trained to represent high-f...

Please sign up or login with your details

Forgot password? Click here to reset