Journal of Computer Applications ›› 2023, Vol. 43 ›› Issue (3): 728-735.DOI: 10.11772/j.issn.1001-9081.2022010034

• Artificial intelligence • Previous Articles    

Video-based person re-identification method based on graph convolution network and self-attention graph pooling

Yingmao YAO, Xiaoyan JIANG()   

  1. School of Electronic and Electrical Engineering,Shanghai University of Engineering Science,Shanghai 201620,China
  • Received:2022-01-13 Revised:2022-03-10 Accepted:2022-03-14 Online:2022-05-31 Published:2023-03-10
  • Contact: Xiaoyan JIANG
  • About author:YAO Yingmao, born in 1997, M. S. candidate. His research interests include person re-identification.
    JIANG Xiaoyan, born in 1985, Ph. D. candidate, associate professor. Her research interests include computer vision.
  • Supported by:
    National Natural Science Foundation of China(U2033218)

基于图卷积网络与自注意力图池化的视频行人重识别方法

姚英茂, 姜晓燕()   

  1. 上海工程技术大学 电子电气工程学院,上海 201620
  • 通讯作者: 姜晓燕
  • 作者简介:姚英茂(1997—),男,河南孟州人,硕士研究生,主要研究方向:行人重识别
    姜晓燕(1985—),女,江苏南通人,副教授,博士研究生,主要研究方向:计算机视觉。
  • 基金资助:
    国家自然科学基金资助项目(U2033218)

Abstract:

Aiming at the bad effect of video person re-identification caused by factors such as occlusion, spatial misalignment and background clutter in cross-camera network videos, a video-based person re-identification method based on Graph Convolutional Network (GCN) and Self-Attention Graph Pooling (SAGP) was proposed. Firstly, the correlation information of different regions between frames in the video was mined through the patch relation graph modeling.In order to alleviate the problems such as occlusion and misalignment, the region features in the frame-by-frame images were optimized by using GCN. Then, the regions with low contribution to person features were removed by SAGP mechanism to avoid the interference of background clutter regions. Finally, a weighted loss function strategy was proposed, the center loss was used to optimize the classification learning results, and Online soft mining and Class-aware attention Loss (OCL) were used to solve the problem that the available samples were not fully used in the process of hard sample mining. Experimental results on MARS dataset show that compared with the sub-optimal Attribute-aware Identity-hard Triplet Loss (AITL), the proposed method has the mean Average Precision (mAP) and Rank-1 increased by 1.3 percentage points and 2.0 percentage points. The proposed method can better utilize the spatial-temporal information in the video to extract more discriminative person features, and improve the effect of person re-identification tasks.

Key words: video-based person re-identification, Graph Convolutional Network (GCN), Self-Attention Graph Pooling (SAGP), weighted loss function strategy, center loss

摘要:

针对跨相机网络视频中存在的遮挡、空间不对齐、背景杂波等因素导致视频行人重识别效果较差的问题,提出一种基于图卷积网络(GCN)与自注意力图池化(SAGP)的视频行人重识别方法。首先,通过区块关系图建模挖掘视频中帧间不同区域的关联信息,并利用GCN优化逐帧图像中的区域特征,缓解遮挡和不对齐等问题;然后,通过SAGP机制去除对行人特征贡献较低的区域,避免背景杂波区域的干扰;最后,提出一种加权损失函数策略,使用中心损失优化分类学习结果,并使用在线软挖掘和类感知注意力(OCL)损失解决难样本挖掘过程中可用样本未被充分利用的问题。实验结果表明,在MARS数据集上,相较于次优的AITL方法,所提方法的平均精度均值(mAP)与Rank-1分别提高1.3和2.0个百点。所提方法能够较好地利用视频中的时空信息,提取更具判别力的行人特征,提高行人重识别任务的效果。

关键词: 视频行人重识别, 图卷积网络, 自注意力图池化, 加权损失函数策略, 中心损失

CLC Number: