Towards a Theoretical Understanding of the Robustness of Variational Autoencoders

07/14/2020
by   Alexander Camuto, et al.
16

We make inroads into understanding the robustness of Variational Autoencoders (VAEs) to adversarial attacks and other input perturbations. While previous work has developed algorithmic approaches to attacking and defending VAEs, there remains a lack of formalization for what it means for a VAE to be robust. To address this, we develop a novel criterion for robustness in probabilistic models: r-robustness. We then use this to construct the first theoretical results for the robustness of VAEs, deriving margins in the input space for which we can provide guarantees about the resulting reconstruction. Informally, we are able to define a region within which any perturbation will produce a reconstruction that is similar to the original reconstruction. To support our analysis, we show that VAEs trained using disentangling methods not only score well under our robustness metrics, but that the reasons for this can be interpreted through our theoretical results.

READ FULL TEXT

page 9

page 20

page 21

research
03/18/2022

Defending Variational Autoencoders from Adversarial Attacks with MCMC

Variational autoencoders (VAEs) are deep generative models used in vario...
research
03/10/2021

Diagnosing Vulnerability of Variational Auto-Encoders to Adversarial Attacks

In this work, we explore adversarial attacks on the Variational Autoenco...
research
02/15/2021

Certifiably Robust Variational Autoencoders

We introduce an approach for training Variational Autoencoders (VAEs) th...
research
02/23/2022

On PAC-Bayesian reconstruction guarantees for VAEs

Despite its wide use and empirical successes, the theoretical understand...
research
02/04/2019

Theoretical evidence for adversarial robustness through randomization: the case of the Exponential family

This paper investigates the theory of robustness against adversarial att...
research
12/04/2020

Unsupervised Adversarially-Robust Representation Learning on Graphs

Recent works have demonstrated that deep learning on graphs is vulnerabl...
research
05/29/2019

Empirically Measuring Concentration: Fundamental Limits on Intrinsic Robustness

Many recent works have shown that adversarial examples that fool classif...

Please sign up or login with your details

Forgot password? Click here to reset