1 |
CANNATA G, SGORBISSA A. A minimalist algorithm for multirobot continuous coverage[J]. IEEE Transactions on Robotics, 2011, 27(2): 297-312. 10.1109/tro.2011.2104510
|
2 |
PORTUGAL D, ROCHA R P. Multi-robot patrolling algorithms: examining performance and scalability[J]. Advanced Robotics, 2013, 27(5): 325-336. 10.1080/01691864.2013.763722
|
3 |
MACHADO A, RAMALHO G, ZUCKER J D, et al. Multi-agent patrolling: an empirical analysis of alternative architectures[C]// Proceedings of the 2002 International Workshop on Multi-Agent Systems and Agent-Based Simulation, LNCS 2581. Berlin: Springer, 2003: 155-170.
|
4 |
PASQUALETTI F, FRANCHI A, BULLO F. On cooperative patrolling: optimal trajectories, complexity analysis, and approximation algorithms[J]. IEEE Transactions on Robotics, 2012, 28(3): 592-606. 10.1109/tro.2011.2179580
|
5 |
ELMALIACH Y, AGMON N, KAMINKA G A. Multi-robot area patrol under frequency constraints[J]. Annals of Mathematics and Artificial Intelligence, 2009, 57(3/4): 293-320. 10.1007/s10472-010-9193-y
|
6 |
CHEN Y, SHU Y, HU M, et al. Multi-UAV cooperative path planning with monitoring privacy preservation[J]. Applied Sciences, 2022, 12(23): No.12111. 10.3390/app122312111
|
7 |
ZHANG H, ZHAO J, WANG R, et al. Multi-objective reinforcement learning algorithm and its application in drive system[C]// Proceedings of the 34th Annual Conference of IEEE Industrial Electronics. Piscataway: IEEE, 2008: 274-279. 10.1109/iecon.2008.4757965
|
8 |
OH J, GUO X, LEE H, et al. Action-conditional video prediction using deep networks in Atari games[C]// Proceedings of the 28th International Conference on Neural Information Processing Systems — Volume 2. Cambridge: MIT Press, 2015: 2863-2871.
|
9 |
CAICEDO J C, LAZEBNIK S. Active object localization with deep reinforcement learning[C]// Proceedings of the 2015 IEEE International Conference on Computer Vision. Piscataway: IEEE, 2015: 2488-2496. 10.1109/iccv.2015.286
|
10 |
LEWIS M, YARATS D, DAUPHIN Y, et al. Deal or no deal? end-to-end learning of negotiation dialogues[C]// Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA: ACL, 2017: 2443-2453. 10.18653/v1/d17-1259
|
11 |
WEISZ G, BUDZIANOWSKI P, SU P H, et al. Sample efficient deep reinforcement learning for dialogue systems with large action spaces[J]. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2018, 26(11): 2083-2097. 10.1109/taslp.2018.2851664
|
12 |
DERHAMI V, PAKSIMA J, KHAJAH H. Web pages ranking algorithm based on reinforcement learning and user feedback[J]. Journal of AI and Data Mining, 2015, 3(2): 157-168. 10.5829/idosi.jaidm.2015.03.02.05
|
13 |
BELLMAN R. On the theory of dynamic programming[J]. Proceedings of the National Academy of Sciences of the United States of America, 1952, 38(8): 716-719. 10.1073/pnas.38.8.716
|
14 |
BRAVO R Z B, LEIRAS A, CYRINO OLIVEIRA F L. The use of UAV s in humanitarian relief: an application of POMDP-based methodology for finding victims[J]. Production and Operations Management, 2019, 28(2): 421-440. 10.1111/poms.12930
|
15 |
BURKS L, AHMED N, LOEFGREN I, et al. Collaborative human-autonomy semantic sensing through structured POMDP planning[J]. Robotics and Autonomous Systems, 2021, 140: No.103753. 10.1016/j.robot.2021.103753
|
16 |
AKBARINASAJI S, KAVAKLIOGLU C, BAŞAR A, et al. Partially observable Markov decision process to generate policies in software defect management[J]. Journal of Systems and Software, 2020, 163: No.110518. 10.1016/j.jss.2020.110518
|
17 |
HORÁK K, BOŠANSKÝ B, PÉCHOUČEK M. Heuristic search value iteration for one-sided partially observable stochastic games[C]// Proceedings of the 31st AAAI Conference on Artificial Intelligence. Palo Alto, CA: AAAI Press, 2017:558-564. 10.1609/aaai.v31i1.10597
|
18 |
LIU F, HUA X, JIN X. A hybrid heuristic value iteration algorithm for POMDP[C]// Proceedings of the IEEE 28th International Conference on Tools with Artificial Intelligence. Piscataway: IEEE, 2016: 304-310. 10.1109/ictai.2016.0054
|
19 |
房俊恒. 基于点的值迭代算法在POMDP问题中的研究[D]. 苏州:苏州大学, 2015: 25-35.
|
|
FANG J H. Research on point-based value iteration algorithms in POMDP domains[D]. Suzhou: Soochow University, 2015: 25-35.
|
20 |
WASHINGTON P H, SCHWAGER M. Reduced state value iteration for multi-drone persistent surveillance with charging constraints[C]// Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems. Piscataway: IEEE, 2021: 6390-6397. 10.1109/iros51168.2021.9636160
|
21 |
BETHKE B, BERTUCCELLI L, HOW J P. Experimental demonstration of adaptive MDP-based planning with model uncertainty[C]// Proceedings of the 2008 AIAA Guidance, Navigation and Control Conference and Exhibit. Reston, VA: AIAA, 2008: No.6322. 10.2514/6.2008-6322
|
22 |
JEONG B M, HA J S, CHOI H L. MDP-based mission planning for multi-UAV persistent surveillance[C]// Proceedings of the 14th International Conference on Control, Automation and Systems. Piscataway: IEEE, 2014: 831-834. 10.1109/iccas.2014.6987894
|
23 |
陈佳,游晓明,刘升,等. 结合信息熵的多种群博弈蚁群算法[J]. 计算机工程与应用, 2019, 55(16):170-178.
|
|
CHEN J, YOU X M, LIU S, et al. Entropy-game based multi-population ant colony optimization[J]. Computer Engineering and Applications, 2019, 55(16):170-178.
|
24 |
HA M, WANG D, LIU D. Generalized value iteration for discounted optimal control with stability analysis[J]. Systems and Control Letters, 2021, 147: No.104847. 10.1016/j.sysconle.2020.104847
|