A Novel Neural Network Training Framework with Data Assimilation

10/06/2020
by   Chong Chen, et al.
0

In recent years, the prosperity of deep learning has revolutionized the Artificial Neural Networks. However, the dependence of gradients and the offline training mechanism in the learning algorithms prevents the ANN for further improvement. In this study, a gradient-free training framework based on data assimilation is proposed to avoid the calculation of gradients. In data assimilation algorithms, the error covariance between the forecasts and observations is used to optimize the parameters. Feedforward Neural Networks (FNNs) are trained by gradient decent, data assimilation algorithms (Ensemble Kalman Filter (EnKF) and Ensemble Smoother with Multiple Data Assimilation (ESMDA)), respectively. ESMDA trains FNN with pre-defined iterations by updating the parameters using all the available observations which can be regard as offline learning. EnKF optimize FNN when new observation available by updating parameters which can be regard as online learning. Two synthetic cases with the regression of a Sine Function and a Mexican Hat function are assumed to validate the effectiveness of the proposed framework. The Root Mean Square Error (RMSE) and coefficient of determination (R2) are used as criteria to assess the performance of different methods. The results show that the proposed training framework performed better than the gradient decent method. The proposed framework provides alternatives for online/offline training the existing ANNs (e.g., Convolutional Neural Networks, Recurrent Neural Networks) without the dependence of gradients.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/01/2023

Sequential Learning from Noisy Data: Data-Assimilation Meets Echo-State Network

This paper explores the problem of training a recurrent neural network f...
research
08/03/2019

Ensemble Neural Networks (ENN): A gradient-free stochastic method

In this study, an efficient stochastic gradient-free method, the ensembl...
research
03/26/2021

A generalised and fully Bayesian framework for ensemble updating

We propose a generalised framework for the updating of a prior ensemble ...
research
07/05/2019

A Unified Framework of Online Learning Algorithms for Training Recurrent Neural Networks

We present a framework for compactly summarizing many recent results in ...
research
05/04/2023

Adjoint-Free 4D-Var Methods Via Line Search Optimization For Non-Linear Data Assimilation

This paper proposes two practical implementations of Four-Dimensional Va...
research
01/25/2021

Finding hidden-feature depending laws inside a data set and classifying it using Neural Network

The logcosh loss function for neural networks has been developed to comb...
research
05/25/2022

On the Interpretability of Regularisation for Neural Networks Through Model Gradient Similarity

Most complex machine learning and modelling techniques are prone to over...

Please sign up or login with your details

Forgot password? Click here to reset