Layerwise Sparsifying Training and Sequential Learning Strategy for Neural Architecture Adaptation

11/13/2022
by   C G Krishnanunni, et al.
0

This work presents a two-stage framework for progressively developing neural architectures to adapt/ generalize well on a given training data set. In the first stage, a manifold-regularized layerwise sparsifying training approach is adopted where a new layer is added each time and trained independently by freezing parameters in the previous layers. In order to constrain the functions that should be learned by each layer, we employ a sparsity regularization term, manifold regularization term and a physics-informed term. We derive the necessary conditions for trainability of a newly added layer and analyze the role of manifold regularization. In the second stage of the Algorithm, a sequential learning process is adopted where a sequence of small networks is employed to extract information from the residual produced in stage I and thereby making robust and more accurate predictions. Numerical investigations with fully connected network on prototype regression problem, and classification problem demonstrate that the proposed approach can outperform adhoc baseline networks. Further, application to physics-informed neural network problems suggests that the method could be employed for creating interpretable hidden layers in a deep network while outperforming equivalent baseline networks.

READ FULL TEXT

page 9

page 16

page 18

page 19

page 20

research
11/22/2015

Gradual DropIn of Layers to Train Very Deep Neural Networks

We introduce the concept of dynamically growing a neural network during ...
research
02/10/2023

Numerical Methods For PDEs Over Manifolds Using Spectral Physics Informed Neural Networks

We introduce an approach for solving PDEs over manifolds using physics i...
research
09/07/2022

Inverse modeling of nonisothermal multiphase poromechanics using physics-informed neural networks

We propose a solution strategy for parameter identification in multiphas...
research
07/03/2019

Spatially-Coupled Neural Network Architectures

In this work, we leverage advances in sparse coding techniques to reduce...
research
11/11/2022

Multilevel-in-Layer Training for Deep Neural Network Regression

A common challenge in regression is that for many problems, the degrees ...
research
08/12/2021

Existence, Stability And Scalability Of Orthogonal Convolutional Neural Networks

Imposing orthogonal transformations between layers of a neural network h...
research
07/16/2021

Entropic alternatives to initialization

Local entropic loss functions provide a versatile framework to define ar...

Please sign up or login with your details

Forgot password? Click here to reset