Tailoring: encoding inductive biases by optimizing unsupervised objectives at prediction time

09/22/2020
by   Ferran Alet, et al.
8

From CNNs to attention mechanisms, encoding inductive biases into neural networks has been a fruitful source of improvement in machine learning. Auxiliary losses are a general way of encoding biases in order to help networks learn better representations by adding extra terms to the loss function. However, since they are minimized on the training data, they suffer from the same generalization gap as regular task losses. Moreover, by changing the loss function, the network is optimizing a different objective than the one we care about. In this work we solve both problems: first, we take inspiration from transductive learning and note that, after receiving an input but before making a prediction, we can fine-tune our models on any unsupervised objective. We call this process tailoring, because we customize the model to each input. Second, we formulate a nested optimization (similar to those in meta-learning) and train our models to perform well on the task loss after adapting to the tailoring loss. The advantages of tailoring and meta-tailoring are discussed theoretically and demonstrated empirically on several diverse examples: encoding inductive conservation laws from physics to improve predictions, improving local smoothness to increase robustness to adversarial examples, and using contrastive losses on the query image to improve generalization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/06/2021

Noether Networks: Meta-Learning Useful Conserved Quantities

Progress in machine learning (ML) stems from a combination of data avail...
research
10/08/2021

Meta-Learning with Task-Adaptive Loss Function for Few-Shot Learning

In few-shot learning scenarios, the challenge is to generalize and perfo...
research
06/04/2019

Hamiltonian Neural Networks

Even though neural networks enjoy widespread use, they still struggle to...
research
06/14/2018

Neural Stethoscopes: Unifying Analytic, Auxiliary and Adversarial Network Probing

Model interpretability and systematic, targeted model adaptation present...
research
12/01/2022

Task Discovery: Finding the Tasks that Neural Networks Generalize on

When developing deep learning models, we usually decide what task we wan...
research
12/14/2020

A case for new neural network smoothness constraints

How sensitive should machine learning models be to input changes? We tac...

Please sign up or login with your details

Forgot password? Click here to reset