《计算机应用》唯一官方网站 ›› 2025, Vol. 45 ›› Issue (10): 3067-3073.DOI: 10.11772/j.issn.1001-9081.2024101535

• 人工智能 •    

基于动态上采样的轻量级生成对抗网络DU-FastGAN

徐国愚, 闫晓龙(), 张一丹   

  1. 河南财经政法大学 计算机与信息工程学院,郑州 450046
  • 收稿日期:2024-10-31 修回日期:2024-12-12 接受日期:2024-12-20 发布日期:2025-03-18 出版日期:2025-10-10
  • 通讯作者: 闫晓龙
  • 作者简介:徐国愚(1982—),男,安徽庐江人,副教授,博士,CCF高级会员,主要研究方向:深度学习
    闫晓龙(2000—),男,河南郑州人,硕士研究生,CCF会员,主要研究方向:深度学习
    张一丹(2001—),女,河南郑州人,硕士研究生,CCF会员,主要研究方向:深度学习。
  • 基金资助:
    国家自然科学基金资助项目(61602153)

DU-FastGAN: lightweight generative adversarial network based on dynamic-upsample

Guoyu XU, Xiaolong YAN(), Yidan ZHANG   

  1. School of Computer and Information Engineering,Henan University of Economics and Law,Zhengzhou Henan 450046,China
  • Received:2024-10-31 Revised:2024-12-12 Accepted:2024-12-20 Online:2025-03-18 Published:2025-10-10
  • Contact: Xiaolong YAN
  • About author:XU Guoyu, born in 1982, Ph. D., associate professor. His research interests include deep learning.
    YAN Xiaolong, born in 2000, M. S. candidate. His research interests include deep learning.
    ZHANG Yidan, born in 2001, M. S. candidate. Her research interests include deep learning.
  • Supported by:
    National Natural Science Foundation of China(61602153)

摘要:

近年来,生成对抗网络(GAN)被广泛应用于数据增强,能有效缓解训练样本不足的问题,对模型训练具有重要研究意义。然而,现有用于数据增强的GAN模型存在对数据集要求高和模型收敛不稳定等问题,导致生成的图像易出现失真和形变。因此,提出一种基于动态上采样的轻量级GAN——DU-FastGAN(Dynamic-Upsample-FastGAN)进行数据增强。首先,通过动态上采样模块构建生成器,使生成器能够根据当前特征图的大小采用不同粒度的上采样方法,从而重建纹理,提高合成的整体结构和局部细节的质量;其次,为了使模型能够更好地获取图像的全局信息流,提出权重信息跳跃连接模块,以减小卷积及池化操作对特征的扰动,提高模型对不同特征的学习能力,使得模型生成图像的细节更逼真;最后,给出特征丢失损失函数,通过计算采样过程中对应特征图之间的相对距离提高模型生成质量。实验结果表明,相较于FastGAN、MixDL(Mixup-based Distance Learning)和RCL-master(Reverse Contrastive Learning-master)等方法,DU-FastGAN在10个小数据集上的FID(Fréchet Inception Distance)的最大降幅达到23.47%,能够有效缓解生成图像的失真和形变问题,并提高了生成图像的质量;同时,DU-FastGAN的模型训练时间在600 min内,实现了轻量级开销。

关键词: 生成对抗网络, 数据增强, 动态上采样, 权重信息跳跃连接, 特征丢失损失

Abstract:

In recent years, Generative Adversarial Networks (GANs) have been widely used for data augmentation, which can solve the problem of insufficient training samples effectively and has important research significance for model training. However, the existing GAN models for data augmentation have problems such as high requirements for datasets and unstable model convergence, which can lead to distortion and deformation of the generated images. Therefore, a lightweight GAN based on dynamic-upsample — DU-FastGAN (Dynamic-Upsample-FastGAN) was proposed for data augmentation. Firstly, a generator was constructed through a dynamic-upsample module, which enables the generator to use upsampling methods of different granularities based on the size of the current feature map, thereby reconstructing textures, and enhancing overall structure and local detail quality of the synthesis. Secondly, in order to enable the model to better obtain global information flow of images, a weight information skip connection module was proposed to reduce the disturbance of convolution and pooling operations on features, thereby improving the model’s learning ability for different features, and making details of the generated images more realistic. Finally, a feature loss function was given to improve the quality of the model generation by calculating relative distance between the corresponding feature maps during the sampling process. Experimental results show that compared with methods such as FastGAN, MixDL (Mixup-based Distance Learning), and RCL-master (Reverse Contrastive Learning-master), DU-FastGAN achieves a maximum reduction of 23.47% in FID (Fréchet Inception Distance) on 10 small datasets, thereby reducing distortion and deformation problems in the generated images effectively, and improving the quality of the generated images. At the same time, DU-FastGAN achieves lightweight overhead with model training time within 600 min.

Key words: Generative Adversarial Network (GAN), data augmentation, dynamic-upsample, weight information skip connection, feature loss function

中图分类号: