Frequency Principle in Deep Learning Beyond Gradient-descent-based Training

01/04/2021
by   Yuheng Ma, et al.
7

Frequency perspective recently makes progress in understanding deep learning. It has been widely verified in both empirical and theoretical studies that deep neural networks (DNNs) often fit the target function from low to high frequency, namely Frequency Principle (F-Principle). F-Principle sheds light on the strength and the weakness of DNNs and inspires a series of subsequent works, including theoretical studies, empirical studies and the design of efficient DNN structures etc. Previous works examine the F-Principle in gradient-descent-based training. It remains unclear whether gradient-descent-based training is a necessary condition for the F-Principle. In this paper, we show that the F-Principle exists stably in the training process of DNNs with non-gradient-descent-based training, including optimization algorithms with gradient information, such as conjugate gradient and BFGS, and algorithms without gradient information, such as Powell's method and Particle Swarm Optimization. These empirical studies show the universality of the F-Principle and provide hints for further study of F-Principle.

READ FULL TEXT

page 6

page 7

page 8

research
01/19/2022

Overview frequency principle/spectral bias in deep learning

Understanding deep learning is increasingly emergent as it penetrates mo...
research
07/28/2020

Deep frequency principle towards understanding why deeper learning is faster

Understanding the effect of depth in deep learning is a critical problem...
research
04/03/2023

Properties and Potential Applications of Random Functional-Linked Types of Neural Networks

Random functional-linked types of neural networks (RFLNNs), e.g., the ex...
research
01/21/2020

Understanding Why Neural Networks Generalize Well Through GSNR of Parameters

As deep neural networks (DNNs) achieve tremendous success across many ap...
research
04/17/2023

Pointwise convergence theorem of generalized mini-batch gradient descent in deep neural network

The theoretical structure of deep neural network (DNN) has been clarifie...
research
10/05/2016

A Novel Representation of Neural Networks

Deep Neural Networks (DNNs) have become very popular for prediction in m...
research
05/12/2022

Feedback Gradient Descent: Efficient and Stable Optimization with Orthogonality for DNNs

The optimization with orthogonality has been shown useful in training de...

Please sign up or login with your details

Forgot password? Click here to reset