Enhancing Robustness of AI Offensive Code Generators via Data Augmentation

06/08/2023
by   Cristina Improta, et al.
0

In this work, we present a method to add perturbations to the code descriptions, i.e., new inputs in natural language (NL) from well-intentioned developers, in the context of security-oriented code, and analyze how and to what extent perturbations affect the performance of AI offensive code generators. Our experiments show that the performance of the code generators is highly affected by perturbations in the NL descriptions. To enhance the robustness of the code generators, we use the method to perform data augmentation, i.e., to increase the variability and diversity of the training data, proving its effectiveness against both perturbed and non-perturbed code descriptions.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset