Differentiable Weight Masks for Domain Transfer

08/26/2023
by   Samar Khanna, et al.
0

One of the major drawbacks of deep learning models for computer vision has been their inability to retain multiple sources of information in a modular fashion. For instance, given a network that has been trained on a source task, we would like to re-train this network on a similar, yet different, target task while maintaining its performance on the source task. Simultaneously, researchers have extensively studied modularization of network weights to localize and identify the set of weights culpable for eliciting the observed performance on a given task. One set of works studies the modularization induced in the weights of a neural network by learning and analysing weight masks. In this work, we combine these fields to study three such weight masking methods and analyse their ability to mitigate "forgetting” on the source task while also allowing for efficient finetuning on the target task. We find that different masking techniques have trade-offs in retaining knowledge in the source task without adversely affecting target task performance.

READ FULL TEXT
research
01/19/2018

Piggyback: Adding Multiple Tasks to a Single, Fixed Network by Learning to Mask

This work presents a method for adding multiple tasks to a single, fixed...
research
06/09/2021

Optimizing Reusable Knowledge for Continual Learning via Metalearning

When learning tasks over time, artificial neural networks suffer from a ...
research
10/05/2020

Are Neural Nets Modular? Inspecting Functional Modularity Through Differentiable Weight Masks

Neural networks (NNs) whose subnetworks implement reusable functions are...
research
09/08/2022

Cross-Modal Knowledge Transfer Without Task-Relevant Source Data

Cost-effective depth and infrared sensors as alternatives to usual RGB s...
research
07/07/2023

Derivative Free Weight-space Ensembling

Recent work suggests that interpolating between the weights of two speci...
research
11/01/2020

An Information-Geometric Distance on the Space of Tasks

This paper computes a distance between tasks modeled as joint distributi...
research
06/11/2019

Learning Selection Masks for Deep Neural Networks

Data have often to be moved between servers and clients during the infer...

Please sign up or login with your details

Forgot password? Click here to reset