《计算机应用》唯一官方网站

• •    下一篇

基于数据增强和标签噪声的快速对抗训练方法

宋逸飞,柳毅   

  1. 广东工业大学
  • 收稿日期:2024-01-02 修回日期:2024-03-14 发布日期:2024-03-28 出版日期:2024-03-28
  • 通讯作者: 宋逸飞
  • 基金资助:
    广东省重点领域研发计划项目

Fast adversarial training method based on data augmentation and label noise

  • Received:2024-01-02 Revised:2024-03-14 Online:2024-03-28 Published:2024-03-28

摘要: 摘 要: 对抗训练是保护分类模型免受对抗性攻击的有效防御方法。然而,由于在训练过程中生成强对抗样本的高成本,可能需要数量级的额外训练时间。为了克服这一限制,基于单步攻击的快速对抗训练已被探索。以往的工作从样本初始化、损失正则化和训练策略等不同角度对快速对抗训练进行了改进。然而,在处理大扰动预算时遇到了灾难性过拟合。基于数据增强与标签噪声的快速对抗训练方法被提出,以解决此困难。初始阶段,对原始样本执行多种图像转换,并引入随机噪声以实施数据增强;接着,少量标签噪声被注入;然后使用增强的数据生成对抗样本用于模型训练;最后,根据对抗鲁棒性测试结果自适应地调整标签噪声率。在CIFAR-10、CIFAR-100数据集上的全面实验结果表明,相较于FGSM-MEP,所提方法在大扰动预算条件下,在两个数据集上的AA上分别提升了4.63和5.38个百分点。经实验证明,新提出的方案可以有效地处理大的扰动预算下灾难性过拟合问题,并显著增强模型的对抗鲁棒性。

关键词: 关键词: 深度学习, 对抗样本, 对抗防御, 数据增强, 标签噪声

Abstract: Abstract: Adversarial training has been an effective defense mechanism for protecting classification models against adversarial attacks. However, the generation of strong adversarial samples during the training process incurred a high computational cost, potentially requiring significantly more training time. To overcome this limitation, fast adversarial training based on single-step attacks was explored. Previous work improved fast adversarial training from different perspectives, such as sample initialization, loss regularization, and training strategies. However, catastrophic overfitting was encountered when dealing with large perturbation budgets. A fast adversarial training method based on data augmentation with label noise is proposed to solve this difficulty. Initially, multiple image transformations are performed on the original samples and random noise is introduced to implement data enhancement; next, a small amount of label noise is injected; then the enhanced data was used to generate adversarial samples for model training; and finally, the label noise rate was adaptively adjusted according to the results of the adversarial robustness test. The comprehensive experimental results on the CIFAR-10 and CIFAR-100 datasets show that compared to FGSM-MEP(Fast Gradient Sign Method with prior from the Momentum of all Previous Epoch), the proposed method improves 4.63 and 5.38 percentage points on the AA(AutoAttack) on the two datasets under the condition of large perturbation budget, respectively. It is experimentally demonstrated that the newly proposed scheme can effectively handle the catastrophic overfitting problem under large perturbation budgets and significantly enhance the adversarial robustness of the model.

Key words: Keywords: deep learning, adversarial example, adversarial defense, data augmentation, label noise

中图分类号: