Towards Optimal Neural Networks: the Role of Sample Splitting in Hyperparameter Selection

07/15/2023
by   Shijin Gong, et al.
0

When artificial neural networks have demonstrated exceptional practical success in a variety of domains, investigations into their theoretical characteristics, such as their approximation power, statistical properties, and generalization performance, have made significant strides. In this paper, we construct a novel theory for understanding the effectiveness of neural networks by discovering the mystery underlying a common practice during neural network model construction: sample splitting. Our theory demonstrates that, the optimal hyperparameters derived from sample splitting can enable a neural network model that asymptotically minimizes the prediction risk. We conduct extensive experiments across different application scenarios and network architectures, and the results manifest our theory's effectiveness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2021

Edge of chaos as a guiding principle for modern neural network training

The success of deep neural networks in real-world problems has prompted ...
research
03/08/2022

Bayesian Optimisation-Assisted Neural Network Training Technique for Radio Localisation

Radio signal-based (indoor) localisation technique is important for IoT ...
research
05/21/2020

HyperSTAR: Task-Aware Hyperparameters for Deep Networks

While deep neural networks excel in solving visual recognition tasks, th...
research
11/22/2020

A Population-based Hybrid Approach to Hyperparameter Optimization for Neural Networks

In recent years, large amounts of data have been generated, and computer...
research
02/09/2019

An Algorithm Unrolling Approach to Deep Image Deblurring

While neural networks have achieved vastly enhanced performance over tra...
research
09/23/2016

Multi-Output Artificial Neural Network for Storm Surge Prediction in North Carolina

During hurricane seasons, emergency managers and other decision makers n...
research
03/23/2020

Steepest Descent Neural Architecture Optimization: Escaping Local Optimum with Signed Neural Splitting

We propose signed splitting steepest descent (S3D), which progressively ...

Please sign up or login with your details

Forgot password? Click here to reset