A Closer Look at the Adversarial Robustness of Information Bottleneck Models

07/12/2021
by   Iryna Korshunova, et al.
0

We study the adversarial robustness of information bottleneck models for classification. Previous works showed that the robustness of models trained with information bottlenecks can improve upon adversarial training. Our evaluation under a diverse range of white-box l_∞ attacks suggests that information bottlenecks alone are not a strong defense strategy, and that previous results were likely influenced by gradient obfuscation.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset