[1] HELBING D, MOLNÁR P. Social force model for pedestrian dynamics[J]. Physical Review E:Statistical Physics, Plasmas, Fluids & Related Interdisciplinary Topics, 1995, 51(5):4282-4286. [2] ROBINETTE P, VELA P A, HOWARD A M. Information propagation applied to robot-assisted evacuation[C]//Proceedings of the 2012 IEEE International Conference on Robotics and Automation. Piscataway:IEEE, 2012:856-861. [3] BOUKAS E, KOSTAVELIS I, GASTERATOS A, et al. Robot guided crowd evacuation[J]. IEEE Transactions on Automation Science and Engineering, 2015, 12(2):739-751. [4] POLYDOROS A S, NALPANTIDIS L. Survey of model-based reinforcement learning:applications on robots[J]. Journal of Intelligent and Robotic Systems, 2017, 86(2):153-173. [5] MNIH V, KAVUKCUOGLU K, SLIVER D, et al. Human-level control through deep reinforcement learning[J]. Nature, 2015, 518(7540):529-533. [6] MNIH V, KAVUKCUOGLU K, SLIVER D, et al. Play Atari with deep reinforcement learning[EB/OL].[2018-12-10]. http://export.arxiv.org/pdf/1312.5602. [7] HWANG K, JIANG W, CHEN Y. Pheromone-based planning strategies in Dyna-Q learning[J]. IEEE Transactions on Industrial Informatics, 2017, 13(2):424-435. [8] IMANBERDIYEV N, FU C, KAYACAN E, et al. Autonomous navigation of UAV by using real-time model-based reinforcement learning[C]//Proceedings of the 14th International Conference on Control, Automation, Robotics and Vision. Piscataway:IEEE, 2016:1-6. [9] GIUSTI A, GUZZI J, CIRESAN D C, et al. A machine learning approach to visual perception of forest trails for mobile robots[J]. IEEE Robotics and Automation Letters, 2016, 1(2):661-667. [10] SU M C, HUANG D, CHOW C, et al. A reinforcement learning approach to robot navigation[C]//Proceedings of the 2004 International Conference on Networking, Sensing and Control. Piscataway:IEEE, 2004:665-669. [11] 胡学敏, 徐珊珊, 康美玉, 等. 基于人机社会力模型的人群疏散算法[J]. 计算机应用, 2018, 38(8):2165-2166. (HU X M, XU S S, KANG M Y, et al. Crowd evacuation based on human-robot social force model[J]. Journal of Computer Applications, 2018, 38(8):2165-2166.) [12] XIE L H, WANG S, MARKHAM A, et al. Towards monocular vision based obstacle avoidance through deep reinforcement learning[EB/OL].[2018-12-10]. https://arxiv.org/pdf/1706.09829.pdf. [13] LILLICRAP T P, HUNT J J, PRITZEL A, et al. Continuous control with deep reinforcement learning[EB/OL].[2019-01-10]. https://arxiv.org/pdf/1509.02971.pdf. [14] CUENCA Á, OJHA U, SALT J, et al. A non-uniform multi-rate control strategy for a Markov chain driven networked control system[J]. Information Sciences, 2015, 321:31-47. [15] 赵玉婷, 韩宝玲, 罗庆生. 基于deep Q-network双足机器人非平整地面行走稳定性控制方法[J]. 计算机应用, 2018, 38(9):2459-2463. (ZHAO Y T, HAN B L, LUO Q S. Walking stability control method based on deep Q-network for biped robot on uneven ground[J]. Journal of Computer Applications, 2018, 38(9):2459-2463.) [16] CHEN Y, LIU M, EVERETT M, et al. Decentralized non-communicating multiagent collision avoidance with deep reinforcement learning[C]//Proceedings of the 2007 IEEE International Conference on Robotics and Automation. Piscataway:IEEE, 2017:285-292. [17] CHEN D, VARSHNEY P K. A survey of void handling techniques or geographic routing in wireless network[J]. IEEE Communications Surveys and Tutorials, 2007, 9(1):50-67. |