Modified Regularized Dual Averaging Method for Training Sparse Convolutional Neural Networks

07/11/2018
by   Xiaodong Jia, et al.
0

We proposed a modified regularized dual averaging method for training sparse deep convolutional neural networks. The regularized dual averaging method has been proven to be effective in obtaining sparse solutions in convex optimization problems, but not applied to deep learning fields before. We analyzed the new version in convex conditions and prove the convergence of it. The modified method can obtain more sparse solutions than traditional sparse optimization methods such as proximal-SGD, while keeping almost the same accuracy as stochastic gradient method with momentum on certain datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/15/2020

Obtaining Adjustable Regularization for Free via Iterate Averaging

Regularization for optimization is a crucial technique to avoid overfitt...
research
10/20/2020

Dual Averaging is Surprisingly Effective for Deep Learning Optimization

First-order stochastic optimization methods are currently the most widel...
research
05/18/2021

Make $\ell_1$ Regularization Effective in Training Sparse CNN

Compressed Sensing using 𝓁1 regularization is among the most powerful an...
research
06/09/2019

Stochastic In-Face Frank-Wolfe Methods for Non-Convex Optimization and Sparse Neural Network Training

The Frank-Wolfe method and its extensions are well-suited for delivering...
research
09/22/2019

A generalization of regularized dual averaging and its dynamics

Excessive computational cost for learning large data and streaming data ...
research
10/30/2019

Unifying mirror descent and dual averaging

We introduce and analyse a new family of algorithms which generalizes an...
research
07/30/2021

Adaptive Optimizers with Sparse Group Lasso for Neural Networks in CTR Prediction

We develop a novel framework that adds the regularizers of the sparse gr...

Please sign up or login with your details

Forgot password? Click here to reset