Fast convergence rates of deep neural networks for classification

12/10/2018
by   Yongdai Kim, et al.
0

We derive the fast convergence rates of a deep neural network (DNN) classifier with the rectified linear unit (ReLU) activation function learned using the hinge loss. We consider three cases for a true model: (1) a smooth decision boundary, (2) smooth conditional class probability, and (3) the margin condition (i.e., the probability of inputs near the decision boundary is small). We show that the DNN classifier learned using the hinge loss achieves fast rate convergences for all three cases provided that the architecture (i.e., the number of layers, number of nodes and sparsity). is carefully selected. An important implication is that DNN architectures are very flexible for use in various cases without much modification. In addition, we consider a DNN classifier learned by minimizing the cross-entropy, and show that the DNN classifier achieves a fast convergence rate under the condition that the conditional class probabilities of most data are sufficiently close to either 1 or zero. This assumption is not unusual for image recognition because human beings are extremely good at recognizing most images. To confirm our theoretical explanation, we present the results of a small numerical study conducted to compare the hinge loss and cross-entropy.

READ FULL TEXT
research
08/02/2021

Convergence rates of deep ReLU networks for multiclass classification

For classification problems, trained deep neural networks return probabi...
research
07/31/2023

Classification with Deep Neural Networks and Logistic Loss

Deep neural networks (DNNs) trained with the logistic loss (i.e., the cr...
research
07/04/2022

Minimax Optimal Deep Neural Network Classifiers Under Smooth Decision Boundary

Deep learning has gained huge empirical successes in large-scale classif...
research
05/22/2023

Multiclass classification for multidimensional functional data through deep neural networks

The intrinsically infinite-dimensional features of the functional observ...
research
03/26/2020

Nonconvex sparse regularization for deep neural networks and its optimality

Recent theoretical studies proved that deep neural network (DNN) estimat...
research
05/24/2023

Fantastic DNN Classifiers and How to Identify them without Data

Current algorithms and architecture can create excellent DNN classifier ...
research
09/28/2020

The model reduction of the Vlasov-Poisson-Fokker-Planck system to the Poisson-Nernst-Planck system via the Deep Neural Network Approach

The model reduction of a mesoscopic kinetic dynamics to a macroscopic co...

Please sign up or login with your details

Forgot password? Click here to reset