Knowledge Evolution in Neural Networks

03/09/2021
by   Ahmed Taha, et al.
0

Deep learning relies on the availability of a large corpus of data (labeled or unlabeled). Thus, one challenging unsettled question is: how to train a deep network on a relatively small dataset? To tackle this question, we propose an evolution-inspired training approach to boost performance on relatively small datasets. The knowledge evolution (KE) approach splits a deep network into two hypotheses: the fit-hypothesis and the reset-hypothesis. We iteratively evolve the knowledge inside the fit-hypothesis by perturbing the reset-hypothesis for multiple generations. This approach not only boosts performance, but also learns a slim network with a smaller inference cost. KE integrates seamlessly with both vanilla and residual convolutional networks. KE reduces both overfitting and the burden for data collection. We evaluate KE on various network architectures and loss functions. We evaluate KE using relatively small datasets (e.g., CUB-200) and randomly initialized deep networks. KE achieves an absolute 21 state-of-the-art baseline. This performance improvement is accompanied by a relative 73 on classification and metric learning benchmarks. Code available at http://bit.ly/3uLgwYb

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/20/2020

Learning the Loss Functions in a Discriminative Space for Video Restoration

With more advanced deep network architectures and learning schemes such ...
research
05/26/2018

Transductive Label Augmentation for Improved Deep Network Learning

A major impediment to the application of deep learning to real-world pro...
research
06/27/2020

ReMarNet: Conjoint Relation and Margin Learning for Small-Sample Image Classification

Despite achieving state-of-the-art performance, deep learning methods ge...
research
12/08/2017

Shape from Shading through Shape Evolution

In this paper, we address the shape-from-shading problem by training dee...
research
09/28/2018

Predicting the Generalization Gap in Deep Networks with Margin Distributions

As shown in recent research, deep neural networks can perfectly fit rand...
research
03/01/2017

Evolving Deep Neural Networks

The success of deep learning depends on finding an architecture to fit t...
research
02/08/2020

On a scalable entropic breaching of the overfitting barrier in machine learning

Overfitting and treatment of "small data" are among the most challenging...

Please sign up or login with your details

Forgot password? Click here to reset