Adversarial Fault Tolerant Training for Deep Neural Networks

07/06/2019
by   Vasisht Duddu, et al.
2

Deep Learning Accelerators are prone to faults which manifest in the form of errors in Neural Networks. Fault Tolerance in Neural Networks is crucial in real-time safety critical applications requiring computation for long durations. Neural Networks with high regularisation exhibit superior fault tolerance, however, at the cost of classification accuracy. In the view of difference in functionality, a Neural Network is modelled as two separate networks, i.e, the Feature Extractor with unsupervised learning objective and the Classifier with a supervised learning objective. Traditional approaches of training the entire network using a single supervised learning objective is insufficient to achieve the objectives of the individual components optimally. In this work, a novel multi-criteria objective function, combining unsupervised training of the Feature Extractor followed by supervised tuning with Classifier Network is proposed. The unsupervised training solves two games simultaneously in the presence of adversary neural networks with conflicting objectives to the Feature Extractor. The first game minimises the loss in reconstructing the input image for indistinguishability given the features from the Extractor, in the presence of a generative decoder. The second game solves a minimax constraint optimisation for distributional smoothening of feature space to match a prior distribution, in the presence of a Discriminator network. The resultant strongly regularised Feature Extractor is combined with the Classifier Network for supervised fine-tuning. The proposed Adversarial Fault Tolerant Neural Network Training is scalable to large networks and is independent of the architecture. The evaluation on benchmarking datasets: FashionMNIST and CIFAR10, indicates that the resultant networks have high accuracy with superior tolerance to stuck at "0" faults compared to widely used regularisers.

READ FULL TEXT
research
11/13/2018

Measuring Masking Fault-Tolerance

In this paper we introduce a notion of fault-tolerance distance between ...
research
11/11/2020

FAT: Training Neural Networks for Reliable Inference Under Hardware Faults

Deep neural networks (DNNs) are state-of-the-art algorithms for multiple...
research
01/12/2023

Exposing Reliability Degradation and Mitigation in Approximate DNNs under Permanent Faults

Approximate computing is known for enhancing deep neural network acceler...
research
05/30/2018

l0-norm Based Centers Selection for Failure Tolerant RBF Networks

The aim of this paper is to select the RBF neural network centers under ...
research
12/13/2020

Fault Injectors for TensorFlow: Evaluation of the Impact of Random Hardware Faults on Deep CNNs

Today, Deep Learning (DL) enhances almost every industrial sector, inclu...
research
12/13/2019

Design of optical neural networks with component imprecisions

For the benefit of designing scalable, fault resistant optical neural ne...
research
12/04/2021

PreGAN: Preemptive Migration Prediction Network for Proactive Fault-Tolerant Edge Computing

Building a fault-tolerant edge system that can quickly react to node ove...

Please sign up or login with your details

Forgot password? Click here to reset