Journal of Computer Applications ›› 2024, Vol. 44 ›› Issue (6): 1807-1815.DOI: 10.11772/j.issn.1001-9081.2023060774

Special Issue: 人工智能

• Artificial intelligence • Previous Articles     Next Articles

Fast adversarial training method based on random noise and adaptive step size

Jinfu WU, Yi LIU()   

  1. School of Computer Science and Technology,Guangdong University of Technology,Guangzhou Guangdong 510006,China
  • Received:2023-06-19 Revised:2023-08-15 Accepted:2023-08-23 Online:2023-09-11 Published:2024-06-10
  • Contact: Jinfu WU, Yi LIU
  • About author:LIU Yi, born in 1976, Ph. D., professor. His research interests include cloud computing security, internet of things security, mobile computing.
  • Supported by:
    Key Technology Research and Development Program of Guangdong Province(2021B0101200002);Key Science and Technology Project of Nansha District, Guangzhou(2022ZD010)

基于随机噪声和自适应步长的快速对抗训练方法

吴锦富, 柳毅()   

  1. 广东工业大学 计算机学院,广州 510006
  • 通讯作者: 吴锦富,柳毅
  • 作者简介:吴锦富(1998—),男,广东梅州人,硕士研究生,主要研究方向:深度学习、图像分类、对抗样本;
  • 基金资助:
    广东省重点领域研发计划项目(2021B0101200002);广州市南沙区重点领域科技项目(2022ZD010)

Abstract:

Adversarial Training (AT) and its variants have been proven to be the most effective methods for defending against adversarial attacks. However, the process of generating adversarial examples requires extensive computational resources, resulting in low model training efficiency and limited feasibility. On the other hand, Fast AT (Fast-AT) uses single-step adversarial attacks to replace multi-step attacks for accelerating the training process, but its model robustness is much lower than that of multi-step AT methods, and it is susceptible to Catastrophic Overfitting (CO). To address these issues, a Fast-AT method based on random noise and adaptive step size was proposed. Firstly, in each iteration of generating adversarial examples, random noise was added to the original input images for data augmentation. Then, the gradients of each adversarial example during the training process were accumulated, and the step size of the adversarial examples was adaptively adjusted based on the gradient information. Finally, adversarial attacks were performed according to the perturbation step size and gradient information to generate adversarial examples for model training. Various adversarial attacks were conducted on the CIFAR-10 and CIFAR-100 datasets, and compared to N-FGSM (Noise Fast Gradient Sign Method), the proposed method achieved at least a 0.35 percentage point improvement in robust accuracy. The experimental results demonstrate that the proposed method can avoid CO issue in Fast-AT and enhance the robustness of deep learning models.

Key words: deep learning, adversarial example, Adversarial Training (AT), random noise, adaptive attack step size

摘要:

当前对抗训练(AT)及其变体被证明是防御对抗攻击的最有效方法,但生成对抗样本的过程需要庞大的计算资源,导致模型训练效率低、可行性不强;快速AT(Fast-AT)使用单步对抗攻击代替多步对抗攻击加速训练过程,但模型鲁棒性远低于多步AT方法且容易发生灾难性过拟合(CO)。针对这些问题,提出一种基于随机噪声和自适应步长的Fast-AT方法。首先,在生成对抗样本的每次迭代中,通过对原始输入图像添加随机噪声增强数据;其次,累积训练过程中每个对抗样本的梯度,并根据梯度信息自适应地调整对抗样本的扰动步长;最后,根据步长和梯度进行对抗攻击,生成对抗样本用于模型训练。在CIFAR-10、CIFAR-100数据集上进行多种对抗攻击,相较于N-FGSM(Noise Fast Gradient Sign Method),所提方法在鲁棒准确率上取得了至少0.35个百分点的提升。实验结果表明,所提方法能避免Fast-AT中的CO问题,提高深度学习模型的鲁棒性。

关键词: 深度学习, 对抗样本, 对抗训练, 随机噪声, 自适应攻击步长

CLC Number: