An Information-Geometric Distance on the Space of Tasks

11/01/2020
by   Yansong Gao, et al.
23

This paper computes a distance between tasks modeled as joint distributions on data and labels. We develop a stochastic process that transports the marginal on the data of the source task to that of the target task, and simultaneously updates the weights of a classifier initialized on the source task to track this evolving data distribution. The distance between two tasks is defined to be the shortest path on the Riemannian manifold of the conditional distribution of labels given data as the weights evolve. We derive connections of this distance with Rademacher complexity-based generalization bounds; distance between tasks computed using our method can be interpreted as the trajectory in weight space that keeps the generalization gap constant as the task distribution changes from the source to the target. Experiments on image classification datasets show that this task distance helps predict the performance of transfer learning: fine-tuning techniques have an easier time transferring to tasks that are close to each other under our distance.

READ FULL TEXT

page 10

page 11

page 12

page 17

research
06/06/2022

Robust Fine-Tuning of Deep Neural Networks with Hessian-based Generalization Guarantees

We consider transfer learning approaches that fine-tune a pretrained dee...
research
03/09/2021

Transfer Learning Can Outperform the True Prior in Double Descent Regularization

We study a fundamental transfer learning process from source to target l...
research
09/09/2017

Optimal Transport for Deep Joint Transfer Learning

Training a Deep Neural Network (DNN) from scratch requires a large amoun...
research
08/29/2019

Learning to Transfer Learn

We propose a novel framework, learning to transfer learn (L2TL), to impr...
research
02/28/2017

Borrowing Treasures from the Wealthy: Deep Transfer Learning through Selective Joint Fine-tuning

Deep neural networks require a large amount of labeled training data dur...
research
01/17/2022

Transfer Learning in Quantum Parametric Classifiers: An Information-Theoretic Generalization Analysis

A key step in quantum machine learning with classical inputs is the desi...
research
08/26/2023

Differentiable Weight Masks for Domain Transfer

One of the major drawbacks of deep learning models for computer vision h...

Please sign up or login with your details

Forgot password? Click here to reset