Functional Indirection Neural Estimator for Better Out-of-distribution Generalization

10/23/2022
by   Kha Pham, et al.
0

The capacity to achieve out-of-distribution (OOD) generalization is a hallmark of human intelligence and yet remains out of reach for machines. This remarkable capability has been attributed to our abilities to make conceptual abstraction and analogy, and to a mechanism known as indirection, which binds two representations and uses one representation to refer to the other. Inspired by these mechanisms, we hypothesize that OOD generalization may be achieved by performing analogy-making and indirection in the functional space instead of the data space as in current methods. To realize this, we design FINE (Functional Indirection Neural Estimator), a neural framework that learns to compose functions that map data input to output on-the-fly. FINE consists of a backbone network and a trainable semantic memory of basis weight matrices. Upon seeing a new input-output data pair, FINE dynamically constructs the backbone weights by mixing the basis weights. The mixing coefficients are indirectly computed through querying a separate corresponding semantic memory using the data pair. We demonstrate empirically that FINE can strongly improve out-of-distribution generalization on IQ tasks that involve geometric transformations. In particular, we train FINE and competing models on IQ tasks using images from the MNIST, Omniglot and CIFAR100 datasets and test on tasks with unseen image classes from one or different datasets and unseen transformation rules. FINE not only achieves the best performance on all tasks but also is able to adapt to small-scale data scenarios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/28/2017

Trainable back-propagated functional transfer matrices

Connections between nodes of fully connected neural networks are usually...
research
05/19/2023

Domain Generalization Deep Graph Transformation

Graph transformation that predicts graph transition from one mode to ano...
research
12/03/2021

Semantic Map Injected GAN Training for Image-to-Image Translation

Image-to-image translation is the recent trend to transform images from ...
research
11/08/2018

Linear Memory Networks

Recurrent neural networks can learn complex transduction problems that r...
research
02/26/2022

Semantic Supervision: Enabling Generalization over Output Spaces

In this paper, we propose Semantic Supervision (SemSup) - a unified para...
research
10/12/2022

Generalization with Lossy Affordances: Leveraging Broad Offline Data for Learning Visuomotor Tasks

The utilization of broad datasets has proven to be crucial for generaliz...
research
07/27/2021

Pointer Value Retrieval: A new benchmark for understanding the limits of neural network generalization

The successes of deep learning critically rely on the ability of neural ...

Please sign up or login with your details

Forgot password? Click here to reset