Large Deviations for Accelerating Neural Networks Training

03/02/2023
by   Sreelekha Guggilam, et al.
0

Artificial neural networks (ANNs) require tremendous amount of data to train on. However, in classification models, most data features are often similar which can lead to increase in training time without significant improvement in the performance. Thus, we hypothesize that there could be a more efficient way to train an ANN using a better representative sample. For this, we propose the LAD Improved Iterative Training (LIIT), a novel training approach for ANN using large deviations principle to generate and iteratively update training samples in a fast and efficient setting. This is exploratory work with extensive opportunities for future work. The thesis presents this ongoing research work with the following contributions from this study: (1) We propose a novel ANN training method, LIIT, based on the large deviations theory where additional dimensionality reduction is not needed to study high dimensional data. (2) The LIIT approach uses a Modified Training Sample (MTS) that is generated and iteratively updated using a LAD anomaly score based sampling strategy. (3) The MTS sample is designed to be well representative of the training data by including most anomalous of the observations in each class. This ensures distinct patterns and features are learnt with smaller samples. (4) We study the classification performance of the LIIT trained ANNs with traditional batch trained counterparts.

READ FULL TEXT

page 1

page 5

page 23

research
01/10/2019

A mixed model approach to drought prediction using artificial neural networks: Case of an operational drought monitoring environment

Droughts, with their increasing frequency of occurrence, continue to neg...
research
07/12/2017

The detector principle of constructing artificial neural networks as an alternative to the connectionist paradigm

Artificial neural networks (ANN) are inadequate to biological neural net...
research
01/25/2019

Pricing options and computing implied volatilities using neural networks

This paper proposes a data-driven approach, by means of an Artificial Ne...
research
08/04/2020

A non-discriminatory approach to ethical deep learning

Artificial neural networks perform state-of-the-art in an ever-growing n...
research
03/13/2021

Efficient Sparse Artificial Neural Networks

The brain, as the source of inspiration for Artificial Neural Networks (...
research
01/03/2022

A Mixed Integer Programming Approach to Training Dense Neural Networks

Artificial Neural Networks (ANNs) are prevalent machine learning models ...
research
10/19/2022

Efficient, probabilistic analysis of combinatorial neural codes

Artificial and biological neural networks (ANNs and BNNs) can encode inp...

Please sign up or login with your details

Forgot password? Click here to reset