Better Together: Resnet-50 accuracy with 13x fewer parameters and at 3x speed

06/10/2020
by   Utkarsh Nath, et al.
0

Recent research on compressing deep neural networks has focused on reducing the number of parameters. Smaller networks are easier to export and deploy on edge-devices. We introduce Adjoined networks as a training approach that can compress and regularize any CNN-based neural architecture. Our one-shot learning paradigm trains both the original and the smaller networks together. The parameters of the smaller network are shared across both the architectures. For resnet-50 trained on Imagenet, we are able to achieve a 13.7x reduction in the number of parameters and a 3x improvement in inference time without any significant drop in accuracy. For the same architecture on CIFAR-100, we are able to achieve a 99.7x reduction in the number of parameters and a 5x improvement in inference time. On both these datasets, the original network trained in the adjoint fashion gains about 3% in top-1 accuracy as compared to the same network trained in the standard fashion.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2019

Associative Convolutional Layers

Motivated by the necessity for parameter efficiency in distributed machi...
research
01/18/2019

Machine Learning with Clos Networks

We present a new methodology for improving the accuracy of small neural ...
research
03/27/2018

Incremental Training of Deep Convolutional Neural Networks

We propose an incremental training method that partitions the original n...
research
10/02/2019

Towards Unifying Neural Architecture Space Exploration and Generalization

In this paper, we address a fundamental research question of significant...
research
04/05/2018

Learning Strict Identity Mappings in Deep Residual Networks

A family of super deep networks, referred to as residual networks or Res...
research
01/19/2020

Towards More Efficient and Effective Inference: The Joint Decision of Multi-Participants

Existing approaches to improve the performances of convolutional neural ...
research
07/16/2018

Backward Reduction of CNN Models with Information Flow Analysis

This paper proposes backward reduction, an algorithm that explores the c...

Please sign up or login with your details

Forgot password? Click here to reset