[1] LAMINI C,FATHI Y,BENHLIMA S.Collaborative Q-learning path planning for autonomous robots based on holonic multi-agent system[C]//Proceedings of the 10th International Conference on Intelligent Systems:Theories and Applications.Washington D.C.,USA:IEEE Press,2015:1-6. [2] HAJDUK M,SUKOP M,HAUN M.Agent approach to multi-agent systems[M]//HAJDUK M,SUKOP M,HAUN M.Studies in systems,decision and control.Berlin,Germany:Springer,2018:21-22. [3] HAN Xiangmin,BAO Hong,LIANG Jun,et al.An adaptive cruise control algorithm based on deep reinforcement learning[J].Computing Engineering,2018,44(7):32-35.(in Chinese)韩向敏,鲍泓,梁军,等.一种基于深度强化学习的自适应巡航控制算法[J].计算机工程,2018,44(7):32-35. [4] YU H L,MEIER K,ARGYLE M,et al.Cooperative path planning for target tracking in urban environments using unmanned air and ground vehicles[J].IEEE/ASME Transactions on Mechatronics,2015,20(2):541-552. [5] YANG P,TANG K,LOZANO J A,et al.Path planning for single unmanned aerial vehicle by separately evolving waypoints[J].IEEE Transactions on Robotics,2015,31(5):1130-1146. [6] ZHOU Hailing,KONG Hui,WEI Lei,et al.Efficient road detection and tracking for unmanned aerial vehicle[J].IEEE Transactions on Intelligent Transportation Systems,2015,16(1):297-309. [7] SHAH K,SCHWAGER M.Multi-agent cooperative pursuit-evasion strategies under uncertainty[M]//CORRELL N,SCHWAGER M,OTTE M.Distributed autonomous robotic systems.Berlin,Germany:Springer,2019:451-468. [8] GARCIA-FERNANDEZ A F,SVENSSON L.Multiple target tracking based on sets of trajectories[J].IEEE Transactions on Aerospace and Electronic Systems,2020,56(3):1685-1707. [9] SOUIDI M E H S,SIAM A,PEI Z Y,et al.Multi-agent pursuit-evasion game based on organizational architecture[J].Journal of Computing and Information Technology,2019,27(1):1-11. [10] DUAN Yong,XU Xinhe.Research on multi-robot cooperation strategy based on multi-agent reinforcement learning[J].Systems Engineering-Theory & Practice,2014,34(5):1305-1310.(in Chinese)段勇,徐心和.基于多智能体强化学习的多机器人协作策略研究[J].系统工程理论与实践,2014,34(5):1305-1310. [11] GUPTA J K,EGOROV M,KOCHENDERFER M.Cooperative multi-agent control using deep reinforcement learning[C]//Proceedings of International Conference on Autonomous Agents and Multiagent Systems.Berlin,Germany:Springer,2017:66-83. [12] WEI E,WICKE D,FREELAN D,et al.Multiagent soft Q-learning[C]//Proceedings of 2018 AAAI Spring Symposium Series.Palo Alto,USA:AAAI Press,2018:1-10. [13] FOERSTER J,FARQUHAR G,AFOURAS T,et al.Counterfactual multi-agent policy gradients[EB/OL].[2019-08-01].https://arxiv.org/abs/1705.08926. [14] LILLICRAP T,HUNT J,PRITZEL A,et al.Continuous control with deep reinforcement learning[EB/OL].[2019-08-01].https://arxiv.org/abs/1509.02971. [15] YAN Yalin.Research on multi-robot pursuit-evasion problem based on game theory[D].Harbin:Harbin Engineering University,2014.(in Chinese)晏亚林.基于博弈论的多机器人追捕问题的研究[D].哈尔滨:哈尔滨工程大学,2014. [16] FANG Baofu,PAN Qishu,HONG Bingrong,et al.Constraint conditions of successful capture in multi-pursuers vs one-evader games[J].Robot,2012,34(3):282-291.(in Chinese)方宝富,潘启树,洪炳镕,等.多追捕者-单-逃跑者追逃问题实现成功捕获的约束条件[J].机器人,2012,34(3):282-291. [17] ZHANG Xu,LI Ling,JIA Leilei.Research and simulation of multi-robot pursuit and escape strategy based on differential game[J].Equipment Manufacturing Technology,2015(9):9-12.(in Chinese)张旭,李玲,贾磊磊.基于微分博弈的多机器人追逃策略研究及仿真[J].装备制造技术,2015(9):9-12. [18] DU Wei,DING Shifei.Overview on multi-agent reinforcement learning[J].Computer Science,2019,46(8):1-8.(in Chinese)杜威,丁世飞.多智能体强化学习综述[J].计算机科学,2019,46(8):1-8. [19] ZHANG Yue.Research on multi-agent deep reinforcement learning methods and applications[D].Xi'an:Xidian University,2018.(in Chinese)张悦.多智能体深度强化学习方法及应用研究[D].西安:西安电子科技大学,2018. [20] WANG Weixun,HAO Jianye,WANG Yixi,et al.Towards cooperation in sequential prisoner's dilemmas:a deep multiagent reinforcement learning approach[EB/OL].[2019-08-01].https://arxiv.org/abs/1803.00162. [21] LOWE R,WU Y,TAMAR A,et al.Multi-agent actor-critic for mixed cooperative-competitive environments[C]//Proceedings of Advances in Neural Information Processing Systems.Berlin,Germany:Springer,2017:6379-6390. |