| 1 |
王洪斌, 尹鹏衡, 郑维, 等. 基于改进的A * 算法与动态窗口法的移动机器人路径规划. 机器人, 2020, 42 (3): 346- 353.
|
|
WANG H B, YIN P H, ZHENG W, et al. Mobile robot path planning based on improved A * algorithm and dynamic window method. Robot, 2020, 42 (3): 346- 353.
|
| 2 |
薛光辉, 王梓杰, 王一凡, 等. 基于改进人工势场算法的煤矿井下机器人路径规划. 工矿自动化, 2024, 50 (5): 6- 13.
|
|
XUE G H, WANG Z J, WANG Y F, et al. Path planning of coal mine underground robot based on improved artificial potential field algorithm. Journal of Mine Automation, 2024, 50 (5): 6- 13.
|
| 3 |
王潇洒, 刘丽星, 杨欣, 等. 改进遗传算法的果园割草机作业路径规划. 重庆理工大学学报(自然科学), 2024, 38 (6): 227- 233.
|
|
WANG X S, LIU L X, YANG X, et al. Improved genetic algorithm for orchard lawn mower operation path planning. Journal of Chongqing University of Technology (Natural Science), 2024, 38 (6): 227- 233.
|
| 4 |
朱敏, 胡若海, 卞京. 基于改进蚁群算法的移动机器人路径规划. 现代制造工程, 2024, 52 (3): 38- 44.
|
|
ZHU M, HU R H, BIAN J. Path planning for mobile robots based on improved ant colony algorithm. Modern Manufacturing Engineering, 2024, 52 (3): 38- 44.
|
| 5 |
HAMZA A. Deep reinforcement learning for mapless mobile robot navigation[D]. Luleå, Sweden: Luleå University of Technology, 2022.
|
| 6 |
JIANG H, WANG H, YUAN W Y, et al. A brief survey: deep reinforcement learning in mobile robot navigation[C]//Proceedings of the 15th IEEE Conference on Industrial Electronics and Applications. Washington D. C., USA: IEEE Press, 2020: 592-597.
|
| 7 |
刘红伟, 潘灵, 吴明钦, 等. 一种FPGA集群轻量级深度学习计算架构设计及实现. 电讯技术, 2024, 64 (1): 14- 21.
|
|
LIU H W, PAN L, WU M Q, et al. Design and implementation of lightweight deep learning computing architecture for FPGA cluster. Telecommunication Engineering, 2024, 64 (1): 14- 21.
|
| 8 |
|
| 9 |
SILVER D, LEVER G, HEESS N, et al. Deterministic policy gradient algorithms[C]//Proceedings of the 31st International Conference on Machine Learning. Washington D. C., USA: IEEE Press, 2014: 387-395.
|
| 10 |
FUJIMOTO S, HOOF H, MEGER D. Addressing function approximation error in actor-critic methods[C]//Proceedings of the 35th International Conference on Machine Learning. Stockholm, Sweden: [s. n.], 2018: 1587-1596.
|
| 11 |
CIMURS R , SUH I H , LEE J H . Goal-driven autonomous exploration through deep reinforcement learning. IEEE Robotics and Automation Letters, 2021, 7 (2): 730- 737.
|
| 12 |
HU W , ZHOU Y , HO H W . Mobile robot navigation based on noisy n-step dueling double deep Q-network and prioritized experience replay. Electronics, 2024, 13 (12): 2423.
doi: 10.3390/electronics13122423
|
| 13 |
SCHAUL T, QUAN J, ANTONOGLOU I, et al. Prioritized experience replay[C]//Proceedings of the 2016 International Conference on Learning Representations. Washington D. C., USA: IEEE Press, 2016: 322-355.
|
| 14 |
MARCHESINI E, FARINELLI A. Discrete deep reinforcement learning for mapless navigation[C]//Proceedings of the 2020 IEEE International Conference on Robotics and Automation. Washington D. C., USA: IEEE Press, 2020: 10688-10694.
|
| 15 |
SAGLAM B , MUTLU F B , CICEK D C , et al. Actor prioritized experience replay. Journal of Artificial Intelligence Research, 2023, 78, 639- 672.
doi: 10.1613/jair.1.14819
|
| 16 |
ZHANG X , SHI X , ZHANG Z , et al. A DDQN path planning algorithm based on experience classification and multi steps for mobile robots. Electronics, 2022, 11 (14): 2120.
doi: 10.3390/electronics11142120
|
| 17 |
VAN HASSELT H, GUEZ A, SILVER D. Deep reinforcement learning with double Q-learning[C]//Proceedings of the 30th AAAI Conference on Artificial Intelligence. Phoenix, USA: AAAI Press, 2016: 2094-2100.
|
| 18 |
YANG J , NI J , LI Y , et al. The intelligent path planning system of agricultural robot via reinforcement learning. Sensors, 2022, 22 (12): 4316.
doi: 10.3390/s22124316
|
| 19 |
HAARNOJA T, ZHOU A, ABBEEL P, et al. Soft Actor-Critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor[C]//Proceedings of the 35th International Conference on Machine Learning. New, York, USA: ACM Press, 2018: 1861-1870.
|
| 20 |
CHOWDHURY M A, LU Q. A novel entropy-maximizing TD3-based reinforcement learning for automatic PID tuning[C]//Proceedings of the 2023 American Control Conference. Washington D. C., USA: IEEE Press, 2023: 2763-2768.
|
| 21 |
YIN Y , CHEN Z , LIU G , et al. A mapless local path planning approach using deep reinforcement learning framework. Sensors, 2023, 23 (4): 2036.
doi: 10.3390/s23042036
|
| 22 |
LIU L , CHEN J , ZHANG Y , et al. Unmanned ground vehicle path planning based on improved DRL algorithm. Electronics, 2024, 13 (13): 2479.
doi: 10.3390/electronics13132479
|
| 23 |
EBERHARD O, HOLLENSTEIN J, PINNERI C, et al. Pink noise is all you need: colored noise exploration in deep reinforcement learning[C]//Proceedings of the 12th International Conference on Learning Representations. Washington D. C., USA: IEEE Press, 2023: 457-466.
|
| 24 |
王涛, 张卫华, 蒲亦非. 适用于强化学习惯性环境的分数阶改进OU噪声. 四川大学学报(自然科学版), 2023, 60 (2): 57- 63.
|
|
WANG T, ZHANG W H, PU Y F. An improved Ornstein-Uhlenbeck exploration noise based on fractional order calculus for reinforcement learning environments with momentum. Journal of Sichuan University (Natural Science Edition), 2023, 60 (2): 57- 63.
|
| 25 |
SUTTON R S , BARTO A G . Reinforcement learning: an introduction. Cambridge, USA: MIT Press, 2018.
|
| 26 |
HESSEL M, MODAYIL J, VAN HASSELT H, et al. Rainbow: combining improvements in deep reinforcement learning[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence. Palo Alto, USA: AAAI Press, 2018: 332-343.
|
| 27 |
FUJIMOTO S, MEGER D, PRECUP D. An equivalence between loss functions and non-uniform sampling in experience replay[C]//Proceedings of the Advances in Neural Information Processing Systems. Cambridge, USA: MIT Press, 2020: 14219-14230.
|
| 28 |
KONDA V, TSITSIKLIS J. Actor-Critic algorithms[C]//Proceedings of the Advances in Neural Information Processing Systems. Cambridge, USA: MIT Press, 1999: 212-222.
|