Adversarial Parameter Attack on Deep Neural Networks

03/20/2022
by   Lijia Yu, et al.
0

In this paper, a new parameter perturbation attack on DNNs, called adversarial parameter attack, is proposed, in which small perturbations to the parameters of the DNN are made such that the accuracy of the attacked DNN does not decrease much, but its robustness becomes much lower. The adversarial parameter attack is stronger than previous parameter perturbation attacks in that the attack is more difficult to be recognized by users and the attacked DNN gives a wrong label for any modified sample input with high probability. The existence of adversarial parameters is proved. For a DNN F_Θ with the parameter set Θ satisfying certain conditions, it is shown that if the depth of the DNN is sufficiently large, then there exists an adversarial parameter set Θ_a for Θ such that the accuracy of F_Θ_a is equal to that of F_Θ, but the robustness measure of F_Θ_a is smaller than any given bound. An effective training algorithm is given to compute adversarial parameters and numerical experiments are used to demonstrate that the algorithms are effective to produce high quality adversarial parameters.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset