Multi-Grade Deep Learning

02/01/2023
by   Yuesheng Xu, et al.
0

The current deep learning model is of a single-grade, that is, it learns a deep neural network by solving a single nonconvex optimization problem. When the layer number of the neural network is large, it is computationally challenging to carry out such a task efficiently. Inspired by the human education process which arranges learning in grades, we propose a multi-grade learning model: We successively solve a number of optimization problems of small sizes, which are organized in grades, to learn a shallow neural network for each grade. Specifically, the current grade is to learn the leftover from the previous grade. In each of the grades, we learn a shallow neural network stacked on the top of the neural network, learned in the previous grades, which remains unchanged in training of the current and future grades. By dividing the task of learning a deep neural network into learning several shallow neural networks, one can alleviate the severity of the nonconvexity of the original optimization problem of a large size. When all grades of the learning are completed, the final neural network learned is a stair-shape neural network, which is the superposition of networks learned from all grades. Such a model enables us to learn a deep neural network much more effectively and efficiently. Moreover, multi-grade learning naturally leads to adaptive learning. We prove that in the context of function approximation if the neural network generated by a new grade is nontrivial, the optimal error of the grade is strictly reduced from the optimal error of the previous grade. Furthermore, we provide several proof-of-concept numerical examples which demonstrate that the proposed multi-grade model outperforms significantly the traditional single-grade model and is much more robust than the traditional model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/13/2023

Successive Affine Learning for Deep Neural Networks

This paper introduces a successive affine learning (SAL) model for const...
research
02/15/2019

Efficient Deep Learning of GMMs

We show that a collection of Gaussian mixture models (GMMs) in R^n can b...
research
09/14/2023

Multi-Grade Deep Learning for Partial Differential Equations with Applications to the Burgers Equation

We develop in this paper a multi-grade deep learning method for solving ...
research
07/17/2018

Are Efficient Deep Representations Learnable?

Many theories of deep learning have shown that a deep network can requir...
research
05/19/2018

GADAM: Genetic-Evolutionary ADAM for Deep Neural Network Optimization

Deep neural network learning can be formulated as a non-convex optimizat...
research
06/18/2021

An Empirical Investigation into Deep and Shallow Rule Learning

Inductive rule learning is arguably among the most traditional paradigms...
research
05/22/2017

Boosting the accuracy of multi-spectral image pan-sharpening by learning a deep residual network

In the field of fusing multi-spectral and panchromatic images (Pan-sharp...

Please sign up or login with your details

Forgot password? Click here to reset