On Model Stability as a Function of Random Seed

09/23/2019
by   Pranava Madhyastha, et al.
0

In this paper, we focus on quantifying model stability as a function of random seed by investigating the effects of the induced randomness on model performance and the robustness of the model in general. We specifically perform a controlled study on the effect of random seeds on the behaviour of attention, gradient-based and surrogate model based (LIME) interpretations. Our analysis suggests that random seeds can adversely affect the consistency of models resulting in counterfactual interpretations. We propose a technique called Aggressive Stochastic Weight Averaging (ASWA)and an extension called Norm-filtered Aggressive Stochastic Weight Averaging (NASWA) which improves the stability of models over random seeds. With our ASWA and NASWA based optimization, we are able to improve the robustness of the original model, on average reducing the standard deviation of the model's performance by 72

READ FULL TEXT
research
11/18/2021

How Emotionally Stable is ALBERT? Testing Robustness with Stochastic Weight Averaging on a Sentiment Analysis Task

Despite their success, modern language models are fragile. Even small ch...
research
02/17/2021

Deviation inequalities for stochastic approximation by averaging

We introduce a class of Markov chains, that contains the model of stocha...
research
10/21/2019

Bayesian Optimization Allowing for Common Random Numbers

Bayesian optimization is a powerful tool for expensive stochastic black-...
research
10/24/2022

We need to talk about random seeds

Modern neural network libraries all take as a hyperparameter a random se...
research
04/21/2020

Robust Motion Averaging under MaximumCorrentropy Criterion

Recently, the motion averaging method has been introduced as an effectiv...
research
06/08/2021

On the Lack of Robust Interpretability of Neural Text Classifiers

With the ever-increasing complexity of neural language models, practitio...
research
09/21/2023

Soft Merging: A Flexible and Robust Soft Model Merging Approach for Enhanced Neural Network Performance

Stochastic Gradient Descent (SGD), a widely used optimization algorithm ...

Please sign up or login with your details

Forgot password? Click here to reset