Stability, Generalization and Privacy: Precise Analysis for Random and NTK Features

05/20/2023
by   Simone Bombari, et al.
0

Deep learning models can be vulnerable to recovery attacks, raising privacy concerns to users, and widespread algorithms such as empirical risk minimization (ERM) often do not directly enforce safety guarantees. In this paper, we study the safety of ERM-trained models against a family of powerful black-box attacks. Our analysis quantifies this safety via two separate terms: (i) the model stability with respect to individual training samples, and (ii) the feature alignment between the attacker query and the original data. While the first term is well established in learning theory and it is connected to the generalization error in classical work, the second one is, to the best of our knowledge, novel. Our key technical result provides a precise characterization of the feature alignment for the two prototypical settings of random features (RF) and neural tangent kernel (NTK) regression. This proves that privacy strengthens with an increase in the generalization capability, unveiling also the role of the activation function. Numerical experiments show a behavior in agreement with our theory not only for the RF and NTK models, but also for deep neural networks trained on standard datasets (MNIST, CIFAR-10).

READ FULL TEXT
research
06/08/2023

Re-aligning Shadow Models can Improve White-box Membership Inference Attacks

Machine learning models have been shown to leak sensitive information ab...
research
02/19/2020

Implicit Regularization of Random Feature Models

Random Feature (RF) models are used as efficient parametric approximatio...
research
02/13/2023

Precise Asymptotic Analysis of Deep Random Feature Models

We provide exact asymptotic expressions for the performance of regressio...
research
12/03/2018

Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks

Deep neural networks are susceptible to various inference attacks as the...
research
04/21/2023

Launching a Robust Backdoor Attack under Capability Constrained Scenarios

As deep neural networks continue to be used in critical domains, concern...
research
10/07/2021

The Connection between Out-of-Distribution Generalization and Privacy of ML Models

With the goal of generalizing to out-of-distribution (OOD) data, recent ...
research
08/27/2020

A Precise Performance Analysis of Learning with Random Features

We study the problem of learning an unknown function using random featur...

Please sign up or login with your details

Forgot password? Click here to reset