《计算机应用》唯一官方网站 ›› 2022, Vol. 42 ›› Issue (10): 3025-3032.DOI: 10.11772/j.issn.1001-9081.2021091571

• 人工智能 • 上一篇    

基于注意力机制和元特征二次重加权的小样本目标检测

林润超, 黄荣, 董爱华   

  1. 东华大学 信息科学与技术学院,上海 201620
  • 收稿日期:2021-09-06 修回日期:2022-01-10 接受日期:2022-01-17 发布日期:2022-04-15 出版日期:2022-10-10
  • 通讯作者: 董爱华
  • 作者简介:第一联系人:林润超(1996—),男,四川宜宾人,硕士研究生,主要研究方向:深度学习、图像处理
    黄荣(1985—),男,浙江绍兴人,讲师,博士,主要研究方向:深度学习、图像理解
    董爱华(1970—),女,上海嘉定人,副教授,博士,主要研究方向:纺织服装人工智能。dongaihua@dhu.edu.cn
  • 基金资助:
    国家重点研发计划项目(2019YFC1521300)

Few-shot object detection based on attention mechanism and secondary reweighting of meta-features

Runchao LIN, Rong HUANG, Aihua DONG   

  1. College of Information Science and Technology,Donghua University,Shanghai 201620,China
  • Received:2021-09-06 Revised:2022-01-10 Accepted:2022-01-17 Online:2022-04-15 Published:2022-10-10
  • Contact: Aihua DONG
  • About author:LIN Runchao, born in 1996, M. S. candidate. His research interests include deep learning, image processing.
    HUANG Rong, born in 1985, Ph. D. , lecturer. His research interests include deep learning, image understanding.
    DONG Aihua, born in 1970, Ph. D. , associate professor. Her research interests include artificial intelligence for textile and apparel.
  • Supported by:
    National Key Research and Development Program of China(2019YFC1521300)

摘要:

在基于迁移学习的小样本目标检测任务中,由于缺乏关注图像中待检测目标的注意力机制,所以现有模型对于待检测目标周边背景区域的抑制能力不强,且在迁移学习过程中通常需要对元特征进行微调来实现跨域共享,这将引起元特征偏移,从而导致模型对大样本图像检测能力的下降。针对上述问题,基于注意力机制和元特征二次重加权机制,提出改进的元特征迁移模型Up-YOLOv3。首先,在原始元特征迁移模型Base-YOLOv2中引入基于卷积块注意力模块(CBAM)的注意力机制,使特征提取网络聚焦于图像中的目标区域并关注图像目标类别的细节特征,从而提升模型对小样本图像目标的检测性能;其次,引入基于压缩?激励(SE)的元特征二次重加权模块(SE-SMFR)对大样本图像的元特征进行二次重加权,以获取二次重加权元特征,使模型在提升小样本目标检测性能的同时也能减小大样本图像元特征信息的权重偏移。实验结果表明,在PASCAL VOC2007/2012数据集上,相较于Base-YOLOv2,Up-YOLOv3针对小样本图像检测的平均准确率均值(mAP)提升了2.3~9.1个百分点;相较于原始的基于YOLOv3元特征迁移模型Base-YOLOv3,Up-YOLOv3针对大样本图像的mAP提升了1.8~2.4个百分点。可见,改进后模型对不同类别的大样本图像和小样本图像均具有良好的泛化能力和鲁棒性。

关键词: 小样本目标检测, 元特征迁移, 特征重加权, 注意力机制, 二次重加权

Abstract:

In the few-shot object detection task based on transfer learning, due to the lack of attention mechanism to focus on the object to be detected in the image, the ability of the existing models to suppress the surrounding background area of the object is not strong, and in the process of transfer learning, it is usually necessary to fine-tune the meta-features to achieve cross-domain sharing, which will cause meta-feature shift, and lead to the decline of the model’s ability to detect large-sample images. To solve the above problems, an improved meta-feature transfer model Up-YOLOv3 based on the attention mechanism and the meta-feature secondary reweighting mechanism was proposed. Firstly, the Convolution Block Attention Module (CBAM)-based attention mechanism was introduced in the original meta-feature transfer model Base-YOLOv2, so that the feature extraction network was able to focus on the object area in the image and pay attention to the detailed features of the image object class, thereby improving the model’s detection performance for few-shot image objects. Then, the Squeeze and Excitation-Secondary Meta-Feature Reweighting (SE-SMFR) module was introduced to reweight the meta-features of the large-sample image for the second time in order to obtain the secondary reweighted meta-features, so that the model was not only able to improve the performance of few-shot object detection, but also able to reduce the weight shift of the meta-feature information of the large-sample image. Experimental results on PASCAL VOC2007/2012 dataset show that, compared with Base-YOLOv2, Up-YOLOv3 has the detection mean Average Precision (mAP) for few-shot object images increased by 2.3 to 9.1 percentage points; compared with the original meta-feature transfer model based on YOLOv3 Base-YOLOv3, mAP for large-sample object images increased by 1.8 to 2.4 percentage points. It can be seen that the improved model has good generalization ability and robustness for both large-sample images and few-shot images of different classes.

Key words: few-shot object detection, meta-feature transfer, feature reweighting, attention mechanism, secondary reweighting

中图分类号: