[1] Xu, Hui and Rongfang Yan.Research on sports action
recognition system based on cluster regression and
improved ISA deep network[J].2020, Intell. Fuzzy Syst.,
39, 5871-5881.
[2] 罗会兰,王婵娟,卢飞.视频行为识别综述[J].通信学
报,2018,39(06):169-180.
[3] 罗会兰,童康,孔繁胜.基于深度学习的视频中人体动作
识别进展综述[J].电子学报,2019,47(05):1162-1173.
[4] Liu X. Sports Deep Learning Method Based on Cognitive
Human Behavior Recognition[J].Computational
Intelligence and Neuroscience, 2022, 2022.
[5] Yao, Guangle, Tao Lei, and Jiandan Zhong. A review of
convolutional-neural-network-based action recognition[J].
Pattern Recognition Letters, 2019, 118: 14-22.
[6] Zhang H B, Zhang Y X, Zhong B, et al. A comprehensive
survey of vision-based human action recognition
methods[J]. Sensors, 2019, 19(5): 1005.
[7] 石 跃 祥 , 朱 茂清 . 基 于 骨架动 作 识 别 的 协 作 卷 积
Transformer 网 络 [J]. 电 子 与 信 息 学
报,2023,45(04):1485-1493.
[8] 赵俊男,佘青山,孟明,陈云.基于多流空间注意力图卷积
SRU 网 络 的 骨 架 动 作 识 别 [J]. 电 子 学
报,2022,50(07):1579-1585.
[9] 王辉,宋佳豪,丁铂栩,何鹏,曹俊杰.三角形网格序列表示
的人体动作识别[J].计算机辅助设计与图形学学
报,2022,34(11):1723-1730.
[10] 王洪雁,袁海.基于骨骼及表观特征融合的动作识别方法
[J].通信学报,2022,43(01):138-148.
[11] Lin J, Gan C, Han S. Tsm: Temporal shift module for
efficient video understanding[C].Proceedings of the
IEEE/CVF international conference on computer vision.
2019: 7083-7093.
[12] Majumder S, Kehtarnavaz N. Vision and inertial sensing
fusion for human action recognition: A review[J]. IEEE
Sensors Journal, 2020, 21(3): 2454-2467.
[13] Wang L, Huynh D Q, Koniusz P. A comparative review of
recent kinect-based action recognition algorithms[J]. IEEE
Transactions on Image Processing, 2019, 29: 15-28.
[14] Z. Tu, J. Zhang, H. Li, Y. Chen, J. Yuan. Joint-bone Fusion
Graph Convolutional Network for Semi-supervised
Skeleton Action Recognition. IEEE Transactions on
Multimedia, vol.25, pp.1819-1831, 2023.
[15] Li Z, Gavrilyuk K, Gavves E, et al. Videolstm convolves,
attends and flows for action recognition[J]. Computer
Vision and Image Understanding, 2018, 166: 41-50.
[16] 胡正平, 刁鹏成, 张瑞雪, 等. 3D 多支路聚合轻量网络
视频行为识别算法研究[J]. 电子学报, 2020, 48(7): 1261
–1268.
[17] 谢昭,周义,吴克伟,张顺然.基于时空关注度 LSTM 的行
为识别[J].计算机学报,2021,44(02):261-274.
[18] 张小俊,李辰政,孙凌宇,张明路.基于改进 3D 卷积神经网
络 的 行 为 识 别 [J]. 计 算 机 集 成 制 造 系
统,2019,25(08):2000-2006.
[19] 于明理. 基于三维卷积神经网络的实时视频动作分类
关键技术研究[D].北京邮电大学,2019.
[20] 石祥滨,李怡颖,刘芳,代钦.T-STAM:基于双流时空注意
力机制的端到端的动作识别模型[J].计算机应用研
究,2021,38(04):1235-1239+1276.
[21] Hara K, Kataoka H, Satoh Y. Can spatiotemporal 3d cnns
retrace the history of 2d cnns and
imagenet?[C].Proceedings of the IEEE conference on
Computer Vision and Pattern Recognition. 2018:
6546-6555. [22] Kay W, Carreira J, Simonyan K, et al. The kinetics human
action video dataset[J]. arXiv preprint arXiv:1705.06950,
2017.
[23] 范银行,赵海峰,张少杰.基于 3D 卷积残差网络的人
体 动 作 识 别 算 法 [J]. 计 算 机 应 用 研
究,2020,37(S2) :300-301,304.
[24] Gao Y, Yang F, Yu Q, et al. Three-dimensional porous
Cu@ Cu 2 O aerogels for direct voltammetric sensing of
glucose[J]. Microchimica Acta, 2019, 186: 1-9.
[25] Woo S, Park J, Lee J Y, et al. Cbam: Convolutional block
attention module[C].Proceedings of the European
conference on computer vision (ECCV). 2018: 3-19.
[26] Cai J, Hu J. 3D RANs: 3D residual attention networks for
action recognition[J]. The Visual Computer, 2020, 36:
1261-1270.
[27] 高德勇,康自兵,王松,王阳萍.利用卷积块注意力机制识
别 人 体 动 作 的 方 法 [J]. 西 安 电 子 科 技 大 学 学
报,2022,49(04):144-155+200.
[28] Li Y, Ji B, Shi X, et al. Tea: Temporal excitation and
aggregation for action recognition[C].Proceedings of the
IEEE/CVF conference on computer vision and pattern
recognition. 2020: 909-918.
[29] Hu J, Shen L, Sun G. Squeeze-and-excitation
networks[C].Proceedings of the IEEE conference on
computer vision and pattern recognition. 2018: 7132-7141.
[30] Soomro K, Zamir A R, Shah M. UCF101: A dataset of 101
human actions classes from videos in the wild[J]. arXiv
preprint arXiv:1212.0402, 2012.
[31] Kuehne H, Jhuang H, Garrote E, et al. HMDB: a large
video database for human motion recognition[C].2011
International conference on computer vision. IEEE, 2011:
2556-2563.
[32] Z. Wang, Q. She and A. Smolic, ACTION-Net: Multipath
Excitation for Action Recognition[C], 2021 IEEE/CVF
Conference on Computer Vision and Pattern Recognition
(CVPR), Nashville, TN, USA, 2021,13209-13218.
[33] D. Zhuang, M. Jiang, J. Kong, T. Liu. Spatiotemporal
attention enhanced features fusion network for action
recognition. International Journal of Machine Learning
and Cybernetics, volume 12, pp.823–841, 2021.
[34] Carreira J, Zisserman A. Quo vadis, action recognition? a
new model and the kinetics dataset[C].proceedings of the
IEEE Conference on Computer Vision and Pattern
Recognition. 2017: 6299-6308.
[35] Zhu J, Zhu Z, Zou W. End-to-end video-level
representation learning for action recognition[C].2018 24th
international conference on pattern recognition (ICPR).
IEEE, 2018: 645-650.
[36] Diba A, Fayyaz M, Sharma V, et al. Temporal 3d convnets:
New architecture and transfer learning for video
classification[J]. arXiv preprint arXiv:1711.08200, 2017.
[37] Wang L, Xiong Y, Wang Z, et al. Temporal segment
networks: Towards good practices for deep action
recognition[C].European conference on computer vision.
Springer, Cham, 2016: 20-36.
|