The race to robustness: exploiting fragile models for urban camouflage and the imperative for machine learning security

06/26/2023
by   Harriet Farlow, et al.
0

Adversarial Machine Learning (AML) represents the ability to disrupt Machine Learning (ML) algorithms through a range of methods that broadly exploit the architecture of deep learning optimisation. This paper presents Distributed Adversarial Regions (DAR), a novel method that implements distributed instantiations of computer vision-based AML attack methods that may be used to disguise objects from image recognition in both white and black box settings. We consider the context of object detection models used in urban environments, and benchmark the MobileNetV2, NasNetMobile and DenseNet169 models against a subset of relevant images from the ImageNet dataset. We evaluate optimal parameters (size, number and perturbation method), and compare to state-of-the-art AML techniques that perturb the entire image. We find that DARs can cause a reduction in confidence of 40.4 benefit of not requiring the entire image, or the focal object, to be perturbed. The DAR method is a deliberately simple approach where the intention is to highlight how an adversary with very little skill could attack models that may already be productionised, and to emphasise the fragility of foundational object detection models. We present this as a contribution to the field of ML security as well as AML. This paper contributes a novel adversarial method, an original comparison between DARs and other AML methods, and frames it in a new context - that of urban camouflage and the necessity for ML security and model robustness.

READ FULL TEXT
research
12/18/2021

Does Explainable Machine Learning Uncover the Black Box in Vision Applications?

Machine learning (ML) in general and deep learning (DL) in particular ha...
research
01/17/2023

Vision Based Machine Learning Algorithms for Out-of-Distribution Generalisation

There are many computer vision applications including object segmentatio...
research
08/24/2022

Bugs in the Data: How ImageNet Misrepresents Biodiversity

ImageNet-1k is a dataset often used for benchmarking machine learning (M...
research
01/28/2019

Strong Black-box Adversarial Attacks on Unsupervised Machine Learning Models

Machine Learning (ML) and Deep Learning (DL) models have achieved state-...
research
09/22/2021

Towards practical object detection for weed spraying in precision agriculture

The evolution of smaller, faster processors and cheaper digital storage ...
research
08/08/2009

Side-channel attack on labeling CAPTCHAs

We propose a new scheme of attack on the Microsoft's ASIRRA CAPTCHA whic...
research
08/31/2021

EG-Booster: Explanation-Guided Booster of ML Evasion Attacks

The widespread usage of machine learning (ML) in a myriad of domains has...

Please sign up or login with your details

Forgot password? Click here to reset