计算机应用 ›› 2018, Vol. 38 ›› Issue (6): 1760-1764.DOI: 10.11772/j.issn.1001-9081.2017112805

• 虚拟现实与多媒体计算 • 上一篇    下一篇

基于时空兴趣点和概率潜动态条件随机场模型的 在线行为识别方法

吴亮, 何毅, 梅雪, 刘欢   

  1. 南京工业大学 电气工程与控制科学学院, 南京 211816
  • 收稿日期:2017-11-29 修回日期:2018-01-05 出版日期:2018-06-10 发布日期:2018-06-13
  • 通讯作者: 何毅
  • 作者简介:吴亮(1992-),江苏宿迁人,硕士研究生,主要研究方向:视频序列、图像处理与识别;何毅(1969-),男,江苏扬州人,副教授,博士,主要研究方向:嵌入式系统;梅雪(1975-),女,内蒙古呼伦贝尔人,副教授,博士,主要研究方向:计算机视觉、图像分析与理解、模式识别、机器学习;刘欢(1990-),男,江苏徐州人,硕士研究生,主要研究方向:图像处理。
  • 基金资助:
    江苏省"六大人才高峰"项目(XXRJ-012);江苏省研究生科研与实践创新计划项目(SJCX17_0276)。

Online behavior recognition using space-time interest points and probabilistic latent-dynamic conditional random field model

WU Liang, HE Yi, MEI Xue, LIU Huan   

  1. College of Electrical Engineering and Control Science, Nanjing Tech University, Nanjing Jiangsu 211816
  • Received:2017-11-29 Revised:2018-01-05 Online:2018-06-10 Published:2018-06-13
  • Supported by:
    This work is partially supported by the Six Talent Peaks Project in Jiangsu Province (XXRJ-012), the Postgraduate Research and Practice Innovation Project in Jiangsu Province (SJCX17_0276).

摘要: 针对在线行为连续序列的识别问题以及行为识别模型的稳定性问题,提出一种监控视频中基于概率潜动态条件随机场(PLDCRF)的在线行为识别方法。首先,应用时空兴趣点(STIP)对行为特征进行提取;再利用PLDCRF模型识别室内人体的活动状态。PLDCRF模型融合了隐含状态变量,能够构建姿态序列子结构,可以选取姿态之间的动态特征,并且直接标记出未分割序列;同时也可以正确地标记出行为间的转换过程,从而明显改善了行为识别的效果。隐含条件随机场(HCRF)、潜动态条件随机场(LDCRF)、潜动态条件神经场(LDCNF)以及PLDCRF模型对10种不同动作的识别率比较结果表明,所提PLDCRF模型对连续的行为序列的综合识别能力更强,并且有更好的稳定性。

关键词: 视频监控, 在线行为识别, 时空兴趣点, 概率潜动态条件随机场

Abstract: In order to improve the recognition ability for online behavior continuous sequences and enhance the stability of behavior recognition model, a novel online behavior recognition method based on Probabilistic Latent-Dynamic Conditional Random Field (PLDCRF) from surveillance video was proposed. Firstly, the Space-Time Interest Point (STIP) was used to extract behavior features. Then, the PLDCRF model was applied to identify the activity state of indoor human body. The proposed PLDCRF model incorporates the hidden state variables and can construct the substructure of gesture sequences. It can select the dynamic features of gesture and mark the unsegmented sequences directly. At the same time, it can also mark the conversion process between behaviors correctly to improve the effect of behavior recognition greatly. Compared with Hidden Conditional Random Field (HCRF), Latent-Dynamic Conditional Random Field (LDCRF) and Latent-Dynamic Conditional Neural Field (LDCNF), the recognition rate comparison results of 10 different behaviors show that, the proposed PLDCRF model has a stronger recognition ability for continuous behavior sequences and better stability.

Key words: video surveillance, online behavior recognition, Space-Time Interest Point (STIP), Probabilistic Latent-Dynamic Conditional Random Field (PLDCRF)

中图分类号: