[1] SILVER D,HUANG A,MADDISON C J,et al.Mastering the game of Go with deep neural networks and tree search[J].Nature,2016,529(7587):484-489. [2] VINYALS O,BABUSCHKIN I,CZARNECKI W M,et al.Grandmaster level in StarCraft Ⅱ using multi-agent reinforcement learning[J].Nature,2019,575(7782):350-354. [3] WEN Kaige,YANG Zhaohui.Intersection signal control based on reinforcement learning with CMAC[J].Computer Engineering,2011,37(17):152-154.(in Chinese)温凯歌,杨照辉.基于CMAC强化学习的交叉口信号控制[J].计算机工程,2011,37(17):152-154. [4] BROCKMAN G,CHEUNG V,PETTERSSON L,et al.OpenAI Gym[EB/OL].(2016-06-05)[2020-02-20].https://arxiv.org/pdf/1606.01540.pdf. [5] TODOROV E,EREZ T,TASSA Y.MuJoCo:a physics engine for model-based control[C]//Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems.Washington D.C.,USA:IEEE Press,2012:5026-5033. [6] TASSA Y,DORON Y,MULDAL A,et al.DeepMind control suite[EB/OL].(2018-01-02)[2020-02-20].https://arxiv.org/pdf/1801.00690.pdf. [7] WYMANN B,ESPIÉ E,GUIONNEAU C,et al.TORCS,the open racing car simulator[EB/OL].(2015-03-12)[2020-02-20].http://www.cse.chalmers.se/~chrdimi/papers/torcs.pdf. [8] DUAN Y,CHEN X,HOUTHOOFT R,et al.Benchmarking deep reinforcement learning for continuous control[C]//Proceedings of International Conference on Machine Learning.New York,USA:[s.n.],2016:1329-1338. [9] BEATTIE C,LEIBO J Z,TEPLYASHIN D,et al.DeepMind lab[EB/OL].(2016-12-12)[2020-02-20].https://arxiv.org/pdf/1612.03801.pdf. [10] COUMANS E,BAI Y.PyBullet,a python module for physics simulation for games,robotics and machine learning[EB/OL].(2016-01-13)[2020-02-20].https://github.com/bulletphysics/bullet3. [11] VINYALS O,EWALDS T,BARTUNOV S,et al.StarCraft Ⅱ:a new challenge for reinforcement learning[EB/OL].(2017-08-16)[2020-02-20].https://arxiv.org/pdf/1708. 04782.pdf. [12] VINITSKY E,KREIDIEH A,FLEM L,et al.Benchmarks for reinforcement learning in mixed-autonomy traffic[C]//Proceedings of Conference on Robot Learning.Zurich,Switzerland:[s.n.],2018:399-409. [13] EREZ T,TASSA Y,TODOROV E.Simulation tools for model-based robotics:comparison of Bullet,Havok,MuJoCo,ODE and PhysX[C]//Proceedings of IEEE International Conference on Robotics and Automation.Washington D.C.,USA:IEEE Press,2015:4397-4404. [14] DASARI S,EBERT F,TIAN S,et al.RoboNet:large-scale multi-robot learning[EB/OL].(2019-10-24)[2020-02-20]. https://arxiv.org/pdf/1910.11215.pdf. [15] MANDLEKAR A,ZHU Y,GARG A,et al.RoboTurk:a crowdsourcing platform for robotic skill learning through imitation[EB/OL].(2018-11-07)[2020-02-20].https://arxiv.org/pdf/1811.02790.pdf. [16] MNIH V,KAVUKCUOGLU K,SILVER D,et al.Human-level control through deep reinforcement learning[J].Nature,2015,518(7540):529-533. [17] POZZI M,PRATTICHIZZO D,MALVEZZI M.On-line educational resources on robotics:a review[C]//Proceedings of International Conference on Inclusive Robotics for a Better Society.Pisa,Italy:[s.n.],2018:141-147. [18] Aliyun.Platform of artificial intelligence[EB/OL].[2020-02-20].https://help.aliyun.com/document_detail/114522.html?spm=5176.12674308.1334604.2113pai798a73dboAt17A. [19] TencentCloud.TI-ONE[EB/OL].[2020-02-20].https://cloud.tencent.com/document/product/851/39399. [20] Baidu.AI studio[EB/OL].[2020-02-20].https://ai.baidu. com/ai-doc/AISTUDIO/Tk39ty6ho. [21] LU Zhonghua,HU Tengteng,WANG Yangang,et al.The design and implement of HPC based on Slurm for deep learning[J].E-science Technology & Application,2018,9(2):40-45.(in Chinese)陆忠华,胡腾腾,王彦棡,等.基于Slurm的深度学习高性能计算平台设计及其调度实现技术[J].科研信息化技术与应用,2018,9(2):40-45. [22] CONGOTE J,SEGURA A,KABONGO L,et al.Interactive visualization of volumetric data with WebGL in real-time[C]//Proceedings of the 16th International Conference on 3D Web Technology.Anaheim,USA:[s.n.],2011:137-146. [23] SEFRAOUI O,AISSAOUI M,ELEULDJ M.OpenStack:toward an open-source solution for cloud computing[J].International Journal of Computer Applications,2012,55(3):38-42. |