[1] WANG Heng,ALEXANDER K,SCHMID C,et al.Dense trajectories and motion boundary descriptors for action recognition[J].International Journal of Computer Vision,2013,103(1):60-79. [2] PENG Xiaojiang,WANG Limin,WANG Xingxing,et al.Bag of visual words and fusion methods for action recognition:comprehensive study and good practice[J].Computer Vision and Image Understanding,2016,150:109-125. [3] WANG Heng,SCHMID C.Action recognition with improved trajectories[C]//Proceedings of IEEE International Confe-rence on Computer Vision.Washington D.C.,USA:IEEE Press,2013:3551-3558. [4] SIMONYAN K,ZISSERMAN A.Two-stream convolutional networks for action recognition in videos[J].Neural Information Processing Systems,2014,1(4):568-576. [5] DONAHUE J,HENDRICKS L A,ROHRBACH M,et al.Long-term recurrent convolutional networks for visual recognition and description[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2014,39(4):677-691. [6] DU T,BOURDEV L,FERGUS R,et al.Learning spatiotemporal features with 3D convolutional networks[C]//Proceedings ofIEEE International Conference on Computer Vision.Washington D.C.,USA:IEEE Computer Society,2015:4489-4497. [7] QIU Zhaofan,YAO Ting,MEI Tao.Learning spatio-temporal representation with pseudo-3D residual networks[C]//Proceedings of 2017 IEEE International Conference on Computer Vision.Washington D.C.,USA:IEEE Press,2017:5534-5542. [8] KENSHO H,HIROKATSU K,YUTAKA S.Towards good practice for action recognition with spatiotemporal 3D convolutions[C]//Proceedings of the 24th International Conference on Pattern Recognition.Washington D.C.,USA:IEEE Press,2018:2516-2521. [9] WANG Limin,XIONG Yuanjun,WANG Zhe,et al.Temporal segment networks:towards good practices for deep action recognition[C]//Proceedings of European Conference on Computer Vision.Berlin,Germany:Springer,2016:20-36. [10] FEICHTENHOFER C,PINZ A,WILDES R P.Spatiotemporal multiplier networks for video action recognition[C]//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition.Washington D.C.,USA:IEEE Press,2017:7445-7454. [11] WANG Jingzhong,HU Kai.Human angle fitting based on BP neural network[J].Computer Systems and Applications,2019,28(8):235-240.(in Chinese) 王景中,胡凯.基于BP回归神经网络的人体角度拟合研究[J].计算机系统应用,2019,28(8):235-240. [12] ZHANG Rui,LI Qishen,CHU Jun.Human action recognition algorithm based on 3D convolution neural network[J].Computer Engineering,2019,45(1):259-263.(in Chinese) 张瑞,李其申,储珺.基于3D卷积神经网络的人体动作识别算法[J].计算机工程,2019,45(1):259-263. [13] LI Wei.Analysis of character motion based on single role video[D].Jinan:Shandong University,2018.(in Chinese) 李伟.基于单角色视频的人物运动分析[D].济南:山东大学,2018. [14] ZIVKOVIC Z.Improved adaptive Gaussian mixture model for background subtraction[C]//Proceedings of International Conference on Pattern Recognition.Washington D.C.,USA:IEEE Press,2004:28-31. [15] QU Jingjing,XIN Yunhong.Combined continuous frame difference with background difference method for moving object detection[J].Acta Photonica Sinica,2014,43(7):219-226.(in Chinese) 屈晶晶,辛云宏.连续帧间差分与背景差分相融合的运动目标检测方法[J].光子学报,2014,43(7):219-226. [16] ZHENG Changyan,MEI Wei,WANG Gang.Deep convolutional neural networks for the image recognition of "S-Maneuver" target[J].Fire Control and Command Control,2017,42(4):66-70.(in Chinese)郑昌艳,梅卫,王刚.基于深度卷积神经网络的蛇形机动航迹图像识别[J].火力与指挥控制,2017,42(4):66-70. [17] SERGEY I,CHRISTIAN S.Batch normalization:accelerating deep network training by reducing internal covariate shift[EB/OL].[2018-12-20].https://arxiv.org/pdf/1502.03167.pdf. [18] MAATEN L V D,HINTON G.Visualizing data using t-SNE[J].Journal of Machine Learning Research,2008,9(3):2579-2605. [19] SCHULDT C,LAPTEV I,CAPUTO B.Recognizing human actions:a local SVM approach[C]//Proceedings of International Conference on Pattern Recognition.Washington D.C.,USA:IEEE Press,2004:32-36. [20] RODRIGUEZ M D,AHMED J,SHAH M.Action MACH:a spatio-temporal maximum average correlation height filter for action recognition[C]//Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.Washington D.C.,USA:IEEE Press,2008:1-8. [21] SOOMRO K,ZAMIR A R,SHAH M.UCF101:a dataset of 101 human actions classes from videos in the wild[EB/OL].[2018-12-20].http://export.arxiv.org/pdf/1212.0402. [22] ZHANG Yahong,LI Yujian.Fisher information metric based on stochastic neighbor embedding[J].Journal of Beijing University of Technology,2016,42(6):862-869.(in Chinese)张亚红,李玉鑑.基于费希尔信息度量的随机近邻嵌入算法[J].北京工业大学学报,2016,42(6):862-869. [23] ZHANG Congxuan,CHEN Zhen,WANG Mingrun,et al.Non-local TV-L1 optical flow estimation using the weighted neighboring triangle filtering[J].Journal of Image and Graphics,2017,22(8):1056-1067.(in Chinese)张聪炫,陈震,汪明润,等.非局部加权邻域三角滤波TV-L1光流估计[J].中国图象图形学报,2017,22(8):1056-1067. [24] GUNNAR F.Two-frame motion estimation based on polynomial expansion[C]//Proceedings of the 13th Scandinavian Conference on Image Analysis.Berlin,Germany:Springer,2003:363-370. |