Breaking the Architecture Barrier: A Method for Efficient Knowledge Transfer Across Networks

12/28/2022
by   Maciej A. Czyzewski, et al.
0

Transfer learning is a popular technique for improving the performance of neural networks. However, existing methods are limited to transferring parameters between networks with same architectures. We present a method for transferring parameters between neural networks with different architectures. Our method, called DPIAT, uses dynamic programming to match blocks and layers between architectures and transfer parameters efficiently. Compared to existing parameter prediction and random initialization methods, it significantly improves training efficiency and validation accuracy. In experiments on ImageNet, our method improved validation accuracy by an average of 1.6 times after 50 epochs of training. DPIAT allows both researchers and neural architecture search systems to modify trained networks and reuse knowledge, avoiding the need for retraining from scratch. We also introduce a network architecture similarity measure, enabling users to choose the best source network without any training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2021

Conceptual Expansion Neural Architecture Search (CENAS)

Architecture search optimizes the structure of a neural network for some...
research
01/07/2021

Transfer Learning Between Different Architectures Via Weights Injection

This work presents a naive algorithm for parameter transfer between diff...
research
11/18/2015

Net2Net: Accelerating Learning via Knowledge Transfer

We introduce techniques for rapidly transferring the information stored ...
research
12/02/2019

Neural Predictor for Neural Architecture Search

Neural Architecture Search methods are effective but often use complex a...
research
08/23/2022

FocusFormer: Focusing on What We Need via Architecture Sampler

Vision Transformers (ViTs) have underpinned the recent breakthroughs in ...
research
03/08/2019

Inductive Transfer for Neural Architecture Optimization

The recent advent of automated neural network architecture search led to...
research
01/13/2022

GradMax: Growing Neural Networks using Gradient Information

The architecture and the parameters of neural networks are often optimiz...

Please sign up or login with your details

Forgot password? Click here to reset