1 |
ZHOU C H, GU S D, WEN Y Q, et al. The review unmanned surface vehicle path planning: based on multi-modality constraint. Ocean Engineering, 2020, 200, 107043.
doi: 10.1016/j.oceaneng.2020.107043
|
2 |
张伟民, 张月, 张辉. 基于改进A*算法的煤矿救援机器人路径规划. 煤田地质与勘探, 2022, 50 (12): 185- 193.
URL
|
|
ZHANG W M, ZHANG Y, ZHANG H. Path planning of coal mine rescue robot based on improved A* algorithm. Coal Geology & Exploration, 2022, 50 (12): 185- 193.
URL
|
3 |
陈丹凤, 雷昊, 刘俊朗, 等. 基于强化蚁群算法的机器人路径规划研究. 兵器装备工程学报, 2023, 44 (6): 239-245, 303.
URL
|
|
CHEN D F, LEI H, LIU J L, et al. Research on robot path planning based on reinforced ant colony optimization. Journal of Ordnance Equipment Engineering, 2023, 44 (6): 239-245, 303.
URL
|
4 |
裴莹, 苏山, 付加胜, 等. 一种求解复杂优化问题的快速遗传算法算子. 吉林大学学报(理学版), 2021, 59 (3): 602- 608.
URL
|
|
PEI Y, SU S, FU J S, et al. A fast genetic algorithm operator for solving complex optimization problems. Journal of Jilin University(Science Edition), 2021, 59 (3): 602- 608.
URL
|
5 |
杨思明, 单征, 曹江, 等. 基于模型的强化学习在无人机路径规划中的应用. 计算机工程, 2022, 48 (12): 255-260, 269.
URL
|
|
YANG S M, SHAN Z, CAO J, et al. Application of model-based reinforcement learning in UAV path planning. Computer Engineering, 2022, 48 (12): 255-260, 269.
URL
|
6 |
MATSUO Y, LECUN Y, SAHANI M, et al. Deep learning, reinforcement learning, and world models. Neural Networks, 2022, 152, 267- 275.
doi: 10.1016/j.neunet.2022.03.037
|
7 |
刘潇, 刘书洋, 庄韫恺, 等. 强化学习可解释性基础问题探索和方法综述. 软件学报, 2023, 34 (5): 2300- 2316.
URL
|
|
LIU X, LIU S Y, ZHUANG Y K, et al. Explainable reinforcement learning: basic problems exploration and method survey. Journal of Software, 2023, 34 (5): 2300- 2316.
URL
|
8 |
|
9 |
|
10 |
ZHAO J D, GAN Z G, LIANG J K, et al. Path planning research of a UAV base station searching for disaster victims' location information based on deep reinforcement learning. Entropy, 2022, 24 (12): 1767.
|
11 |
郑帅, 罗飞, 顾春华, 等. 基于双估计器的改进Speedy Q-learning算法. 计算机科学, 2020, 47 (7): 179- 185.
URL
|
|
ZHENG S, LUO F, GU C H, et al. Improved Speedy Q-learning algorithm based on double estimator. Computer Science, 2020, 47 (7): 179- 185.
URL
|
12 |
LIU B Y, YE X B, ZHOU C F, et al. The improved algorithm of deep Q-learning network based on eligibility trace[C]//Proceedings of the 6th International Conference on Control, Automation and Robotics. Washington D. C., USA: IEEE Press, 2020: 230-235.
|
13 |
HU R J, ZHANG Y L. Fast path planning for long-range planetary roving based on a hierarchical framework and deep reinforcement learning. Aerospace, 2022, 9 (2): 101.
|
14 |
刘全, 闫岩, 朱斐, 等. 一种带探索噪音的深度循环Q网络. 计算机学报, 2019, 42 (7): 1588- 1604.
URL
|
|
LIU Q, YAN Y, ZHU F, et al. A deep recurrent Q network with exploratory noise. Chinese Journal of Computers, 2019, 42 (7): 1588- 1604.
URL
|
15 |
LU J J, LIU W X, ZHU Y H, et al. Scheduling mix-flow in SD-DCN based on deep reinforcement learning with private link[C]//Proceedings of the 16th International Conference on Mobility, Sensing and Networking. Washington D. C., USA: IEEE Press, 2021: 395-401.
|
16 |
|
17 |
赵英男, 刘鹏, 赵巍, 等. 深度Q学习的二次主动采样方法. 自动化学报, 2019, 45 (10): 1870- 1882.
URL
|
|
ZHAO Y N, LIU P, ZHAO W, et al. Twice sampling method in deep Q-network. Acta Automatica Sinica, 2019, 45 (10): 1870- 1882.
URL
|
18 |
LÜ L H, ZHANG S J, DING D R, et al. Path planning via an improved DQN-based learning policy. IEEE Access, 2019, 7, 67319- 67330.
|
19 |
LIU Y H, XU Y Z. Free gait planning of hexapod robot based on improved DQN algorithm[C]//Proceedings of the 2nd International Conference on Civil Aviation Safety and Information Technology. Washington D. C., USA: IEEE Press, 2021: 488-491.
|
20 |
LI J X, CHEN Y T, ZHAO X N, et al. An improved DQN path planning algorithm. The Journal of Supercomputing, 2022, 78 (1): 616- 639.
|
21 |
LIU Y L, CHEN Z G, LI Y G, et al. Robot search path planning method based on prioritized deep reinforcement learning. International Journal of Control, Automation and Systems, 2022, 20 (8): 2669- 2680.
|
22 |
ZHANG Y, WANG T B. Applying value-based deep reinforcement learning on KPI time series anomaly detection[C]//Proceedings of the 15th International Conference on Cloud Computing. Washington D. C., USA: IEEE Press, 2022: 197-202.
|
23 |
刘全, 翟建伟, 章宗长, 等. 深度强化学习综述. 计算机学报, 2018, 41 (1): 1- 27.
URL
|
|
LIU Q, ZHAI J W, ZHANG Z C, et al. A survey on deep reinforcement learning. Chinese Journal of Computers, 2018, 41 (1): 1- 27.
URL
|
24 |
马昂, 于艳华, 杨胜利, 等. 基于强化学习的知识图谱综述. 计算机研究与发展, 2022, 59 (8): 1694- 1722.
URL
|
|
MA A, YU Y H, YANG S L, et al. Summary of knowledge map based on reinforcement learning. Journal of Computer Research and Development, 2022, 59 (8): 1694- 1722.
URL
|
25 |
LIN S W, LIU A, WANG J G, et al. A review of path-planning approaches for multiple mobile robots. Machines, 2022, 10 (9): 773.
|
26 |
董永峰, 杨琛, 董瑶, 等. 基于改进的DQN机器人路径规划. 计算机工程与设计, 2021, 42 (2): 552- 558.
URL
|
|
DONG Y F, YANG C, DONG Y, et al. Robot path planning based on improved DQN. Computer Engineering and Design, 2021, 42 (2): 552- 558.
URL
|
27 |
LEE S M, KIM S B. Parallel simulated annealing with a greedy algorithm for Bayesian network structure learning. IEEE Transactions on Knowledge and Data Engineering, 2020, 32 (6): 1157- 1166.
|