Regularization and Optimization strategies in Deep Convolutional Neural Network

12/13/2017
by   Pushparaja Murugan, et al.
0

Convolution Neural Networks, known as ConvNets exceptionally perform well in many complex machine learning tasks. The architecture of ConvNets demands the huge and rich amount of data and involves with a vast number of parameters that leads the learning takes to be computationally expensive, slow convergence towards the global minima, trap in local minima with poor predictions. In some cases, architecture overfits the data and make the architecture difficult to generalise for new samples that were not in the training set samples. To address these limitations, many regularization and optimization strategies are developed for the past few years. Also, studies suggested that these techniques significantly increase the performance of the networks as well as reducing the computational cost. In implementing these techniques, one must thoroughly understand the theoretical concept of how this technique works in increasing the expressive power of the networks. This article is intended to provide the theoretical concepts and mathematical formulation of the most commonly used strategies in developing a ConvNet architecture.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/19/2017

Hyperparameters Optimization in Deep Convolutional Neural Network / Bayesian Approach with Gaussian Process Prior

Convolutional Neural Network is known as ConvNet have been extensively u...
research
11/19/2019

Information-Theoretic Local Minima Characterization and Regularization

Recent advances in deep learning theory have evoked the study of general...
research
10/02/2020

No Spurious Local Minima: on the Optimization Landscapes of Wide and Deep Neural Networks

Empirical studies suggest that wide neural networks are comparably easy ...
research
01/03/2018

Implementation of Deep Convolutional Neural Network in Multi-class Categorical Image Classification

Convolutional Neural Networks has been implemented in many complex machi...
research
02/22/2021

Non-Convex Optimization with Spectral Radius Regularization

We develop a regularization method which finds flat minima during the tr...
research
11/23/2022

Expressibility-Enhancing Strategies for Quantum Neural Networks

Quantum neural networks (QNNs), represented by parameterized quantum cir...
research
06/11/2020

Mixup Training as the Complexity Reduction

Machine learning has achieved remarkable results in recent years due to ...

Please sign up or login with your details

Forgot password? Click here to reset