[1] YU G, YUAN J, LIU Z. Predicting human activities using spatio-temporal structure of interest points[C]//Proceedings of the 20th ACM International Conference on Multimedia. New York: ACM, 2012: 1049-1052. [2] YU T H, KIM T K, CIPOLLA R. Real-time action recognition by spatiotemporal semantic and structural forests[C]//BMVC 2010: Proceedings of the 21st British Machine Vision Conference. Bristol: BMVA, 2010: 1-12. [3] LAPTEV I, LINDEBERG T. Space-time interest points [C]//Proceedings of the 9th IEEE International Conference on Computer Vision. Piscataway, NJ: IEEE, 2003:432-439. [4] DOLLÁR P, RABAUD V, COTTRELL G, et al. Behavior recognition via sparse spatio-temporal features[C]//Proceedings of the 2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance. Piscataway, NJ: IEEE, 2005: 65-72. [5] BURGHOUTS G J, SCHUTTE K. Spatio-temporal layout of human actions for improved bag-of-words action detection[J]. Pattern Recognition Letters, 2013, 34(15): 1861-1869. [6] ZHANG X, CUI J, TIAN L, et al. Local spatio-temporal feature based voting framework for complex human activity detection and localization[C]//Proceedings of the First Asian Conference on Pattern Recognition. Piscataway, NJ: IEEE, 2011: 12-16. [7] LI N, CHENG X, GUO H, et al. A hybrid method for human interaction recognition using spatio-temporal interest points [C]//ICPR 2014: Proceedings of the 22nd International Conference on Pattern Recognition. Piscataway, NJ: IEEE, 2014: 2513-2518. [8] 韩磊, 李君峰, 贾云得. 基于时空单词的两人交互行为识别方法[J]. 计算机学报, 2010, 33(4): 776-784.(HAN L, LI J F, JIA Y D. Human interaction recognition using spatio-temporal words[J]. Chinese Journal of Computers, 2010, 33(4): 776-784.) [9] GAUR U, ZHU Y, SONG B, et al. A "string of feature graphs" model for recognition of complex activities in natural videos[C]//Proceedings of the 2011 IEEE International Conference on Computer Vision. Piscataway, NJ: IEEE, 2011: 2595-2602. [10] PENG X, PENG Q, QIAO Y, et al. Exploring dense trajectory feature and encoding methods for human interaction recognition[C]//ICIMCS 2013: Proceedings of the Fifth International Conference on Internet Multimedia Computing and Service. Piscataway, NJ: IEEE, 2013: 23-27. [11] 王策, 姬晓飞, 李一波. 一种简便的视角无关动作识别方法[J]. 智能系统学报, 2014, 9(5): 577-583.(WANG C, JI X F, LI Y B. Study on a simple view-invariant action recognition method[J]. CAAI Transactions on Intelligent Systems, 2014, 9(5): 577-583.) [12] 王世刚, 孙爱朦, 赵文婷, 等. 基于时空兴趣点的单人行为及交互行为识别[J]. 吉林大学学报(工学版), 2015,45(1): 304-308.(WANG S G, SUN A M, ZHAO W T, et al. Single and interactive human behavior recognition algorithm based on spatio-temporal interest point[J]. Journal of Jilin University (Engineering and Technology Edition), 2015,45(1): 304-308.) [16] KONG Y, LIANG W, DONG Z Y, et al. Recognising human interaction from videos by a discriminative model[J]. IET Computer Vision, 2014, 8(4): 277-286. [17] PATRON-PEREZ A, MARSZALEK M, REID I, et al. Structured learning of human interactions in TV shows[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, 34(12): 2441-2453. [13] GONG H, ZHU S C. Intrackability: characterizing video statistics and pursuing video representations[J]. International Journal of Computer Vision, 2012, 97(3): 255-275. [14] SCOVANNER P, ALI S, SHAH M. A 3-dimensional SIFT descriptor and its application to action recognition[C]//Proceedings of the 15th International Conference on Multimedia. New York: ACM, 2007: 357-360. [15] 朱旭锋, 马彩文, 刘波. 采用改进词袋模型的空中目标自动分类[J]. 红外与激光工程, 2012, 41(5): 1384-1388.(ZHU X F, MA C W, LIU B. Aerial target automatic classification based on improving bag of words model[J]. Infrared and Laser Engineering, 2012, 41(5):1384-1388.) [16] KONG Y, LIANG W, DONG Z Y, et al. Recognising human interaction from videos by a discriminative model[J]. IET Computer Vision, 2014, 8(4): 277-286. [17] PATRON-PEREZ A, MARSZALEK M, REID I, et al. Structured learning of human interactions in TV shows[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, 34(12): 2441-2453. |