A Bayesian Perspective On Training Data Attribution

05/31/2023
by   Elisa Nguyen, et al.
1

Training data attribution (TDA) techniques find influential training data for the model's prediction on the test data of interest. They approximate the impact of down- or up-weighting a particular training sample. While conceptually useful, they are hardly applicable in practice, particularly because of their sensitivity to different model initialisation. In this paper, we introduce a Bayesian perspective on the TDA task, where the learned model is treated as a Bayesian posterior and the TDA estimates as random variables. From this novel viewpoint, we observe that the influence of an individual training sample is often overshadowed by the noise stemming from model initialisation and SGD batch composition. Based on this observation, we argue that TDA can only be reliably used for explaining model predictions that are consistently influenced by certain training data, independent of other noise factors. Our experiments demonstrate the rarity of such noise-independent training-test data pairs but confirm their existence. We recommend that future researchers and practitioners trust TDA estimates only in such cases. Further, we find a disagreement between ground truth and estimated TDA distributions and encourage future work to study this gap. Code is provided at https://github.com/ElisaNguyen/bayesian-tda.

READ FULL TEXT

page 8

page 15

page 16

page 17

page 18

page 19

research
06/03/2023

Training Data Attribution for Diffusion Models

Diffusion models have become increasingly popular for synthesizing high-...
research
01/31/2020

Stable Prediction with Model Misspecification and Agnostic Distribution Shift

For many machine learning algorithms, two main assumptions are required ...
research
04/09/2021

An Empirical Comparison of Instance Attribution Methods for NLP

Widespread adoption of deep models has motivated a pressing need for app...
research
12/09/2022

Training Data Influence Analysis and Estimation: A Survey

Good models require good training data. For overparameterized deep model...
research
02/17/2016

Authorship Attribution Using a Neural Network Language Model

In practice, training language models for individual authors is often ex...
research
04/14/2021

Detection of a rank-one signal with limited training data

In this paper, we reconsider the problem of detecting a matrix-valued ra...
research
08/28/2023

Task-Aware Machine Unlearning and Its Application in Load Forecasting

Data privacy and security have become a non-negligible factor in load fo...

Please sign up or login with your details

Forgot password? Click here to reset