Denoising Noisy Neural Networks: A Bayesian Approach with Compensation

05/22/2021
by   Yulin Shao, et al.
31

Noisy neural networks (NoisyNNs) refer to the inference and training of NNs in the presence of noise. Noise is inherent in most communication and storage systems; hence, NoisyNNs emerge in many new applications, including federated edge learning, where wireless devices collaboratively train a NN over a noisy wireless channel, or when NNs are implemented/stored in an analog storage medium. This paper studies a fundamental problem of NoisyNNs: how to estimate the uncontaminated NN weights from their noisy observations or manifestations. Whereas all prior works relied on the maximum likelihood (ML) estimation to maximize the likelihood function of the estimated NN weights, this paper demonstrates that the ML estimator is in general suboptimal. To overcome the suboptimality of the conventional ML estimator, we put forth an MMSE_pb estimator to minimize a compensated mean squared error (MSE) with a population compensator and a bias compensator. Our approach works well for NoisyNNs arising in both 1) noisy inference, where noise is introduced only in the inference phase on the already-trained NN weights; and 2) noisy training, where noise is introduced over the course of training. Extensive experiments on the CIFAR-10 and SST-2 datasets with different NN architectures verify the significant performance gains of the MMSE_pb estimator over the ML estimator when used to denoise the NoisyNN. For noisy inference, the average gains are up to 156% for a noisy ResNet34 model and 14.7% for a noisy BERT model; for noisy training, the average gains are up to 18.1 dB for a noisy ResNet18 model.

READ FULL TEXT
research
09/08/2021

Bayesian Over-The-Air Computation

Analog over-the-air computation (OAC) is an efficient solution to a clas...
research
11/20/2020

Noisy Neural Network Compression for Analog Storage Devices

Efficient compression and storage of neural network (NN) parameters is c...
research
09/18/2019

Towards a New Understanding of the Training of Neural Networks with Mislabeled Training Data

We investigate the problem of machine learning with mislabeled training ...
research
02/15/2021

Neural Network Compression for Noisy Storage Devices

Compression and efficient storage of neural network (NN) parameters is c...
research
02/26/2021

Federated Edge Learning with Misaligned Over-The-Air Computation

Over-the-air computation (OAC) is a promising technique to realize fast ...
research
02/01/2016

Cluster-Seeking James-Stein Estimators

This paper considers the problem of estimating a high-dimensional vector...
research
08/05/2021

Multi-task Federated Edge Learning (MtFEEL) in Wireless Networks

Federated Learning (FL) has evolved as a promising technique to handle d...

Please sign up or login with your details

Forgot password? Click here to reset