Model Monitoring and Robustness of In-Use Machine Learning Models: Quantifying Data Distribution Shifts Using Population Stability Index

02/01/2023
by   Aria Khademi, et al.
0

Safety goes first. Meeting and maintaining industry safety standards for robustness of artificial intelligence (AI) and machine learning (ML) models require continuous monitoring for faults and performance drops. Deep learning models are widely used in industrial applications, e.g., computer vision, but the susceptibility of their performance to environment changes (e.g., noise) after deployment on the product, are now well-known. A major challenge is detecting data distribution shifts that happen, comparing the following: (i) development stage of AI and ML models, i.e., train/validation/test, to (ii) deployment stage on the product (i.e., even after `testing') in the environment. We focus on a computer vision example related to autonomous driving and aim at detecting shifts that occur as a result of adding noise to images. We use the population stability index (PSI) as a measure of presence and intensity of shift and present results of our empirical experiments showing a promising potential for the PSI. We further discuss multiple aspects of model monitoring and robustness that need to be analyzed simultaneously to achieve robustness for industry safety standards. We propose the need for and the research direction toward categorizations of problem classes and examples where monitoring for robustness is required and present challenges and pointers for future work from a practical perspective.

READ FULL TEXT

page 12

page 13

research
05/24/2023

Non-adversarial Robustness of Deep Learning Methods for Computer Vision

Non-adversarial robustness, also known as natural robustness, is a prope...
research
11/03/2019

Artificial Intelligence Strategies for National Security and Safety Standards

Recent advances in artificial intelligence (AI) have lead to an explosio...
research
04/26/2022

Landing AI on Networks: An equipment vendor viewpoint on Autonomous Driving Networks

The tremendous achievements of Artificial Intelligence (AI) in computer ...
research
07/29/2021

Did the Model Change? Efficiently Assessing Machine Learning API Shifts

Machine learning (ML) prediction APIs are increasingly widely used. An M...
research
10/28/2020

Evaluating Model Robustness to Dataset Shift

As the use of machine learning in safety-critical domains becomes widesp...
research
03/22/2023

Deployment of Image Analysis Algorithms under Prevalence Shifts

Domain gaps are among the most relevant roadblocks in the clinical trans...
research
05/05/2023

All models are local: time to replace external validation with recurrent local validation

External validation is often recommended to ensure the generalizability ...

Please sign up or login with your details

Forgot password? Click here to reset