Vulnerabilities in AI Code Generators: Exploring Targeted Data Poisoning Attacks

08/04/2023
by   Domenico Cotroneo, et al.
0

In this work, we assess the security of AI code generators via data poisoning, i.e., an attack that injects malicious samples into the training data to generate vulnerable code. We poison the training data by injecting increasing amounts of code containing security vulnerabilities and assess the attack's success on different state-of-the-art models for code generation. Our analysis shows that AI code generators are vulnerable to even a small amount of data poisoning. Moreover, the attack does not impact the correctness of code generated by pre-trained models, making it hard to detect.

READ FULL TEXT
research
06/08/2023

Enhancing Robustness of AI Offensive Code Generators via Data Augmentation

In this work, we present a method to add perturbations to the code descr...
research
10/31/2022

Poison Attack and Defense on Deep Source Code Processing Models

In the software engineering community, deep learning (DL) has recently b...
research
05/22/2023

The "code” of Ethics:A Holistic Audit of AI Code Generators

AI-powered programming language generation (PLG) models have gained incr...
research
07/05/2020

You Autocomplete Me: Poisoning Vulnerabilities in Neural Code Completion

Code autocompletion is an integral feature of modern code editors and ID...
research
09/19/2023

Exploring the Dark Side of AI: Advanced Phishing Attack Design and Deployment Using ChatGPT

This paper explores the possibility of using ChatGPT to develop advanced...
research
04/14/2023

PTW: Pivotal Tuning Watermarking for Pre-Trained Image Generators

Deepfakes refer to content synthesized using deep generators, which, whe...
research
05/27/2023

Backdooring Neural Code Search

Reusing off-the-shelf code snippets from online repositories is a common...

Please sign up or login with your details

Forgot password? Click here to reset