[1] 权美香, 朴松昊, 李国. 视觉SLAM综述[J]. 智能系统学报, 2016, 11(6):768-776. (QUAN M X, PIAO S H, LI G. An overview of visual SLAM[J]. CAAI Transactions on Intelligent Systems, 2016, 11(6):768-776.) [2] MUR-ARTAL R, TARDÓS J D. ORB-SLAM2:an open-source SLAM system for monocular, stereo, and RGB-D cameras[J]. IEEE Transactions on Robotics, 2017, 33(5):1255-1262. [3] ENGEL J, SCHÖPS T, CREMERS D. LSD-SLAM:large-scale direct monocular SLAM[C]//Proceedings of the 2014 European Conference on Computer Vision, LNCS 8690. Cham:Springer, 2014:834-849. [4] 方岚, 于凤芹. 去除鬼影及阴影的视觉背景提取运动目标检测算法[J]. 激光与光电子学进展, 2019, 56(13):No. 131002. (FANG L, YU F Q. Moving object detection algorithm based on removed ghost and shadow visual background extractor[J]. Laser and Optoelectronics Progress, 2019, 56(13):No. 131002.) [5] ENDRES F, HESS J, STURM J, et al. 3-D mapping with an RGBD camera[J]. IEEE Transactions on Robotics, 2014, 30(1):177-187. [6] 赵洋, 刘国良, 田国会, 等. 基于深度学习的视觉SLAM综述[J]. 机器人, 2017, 39(6):889-896.(ZHAO Y, LIU G L, TIAN G H, et al. A survey of visual SLAM based on deep learning[J]. Robot, 2017, 39(6):889-896.) [7] SALAS-MORENO R F, NEWCOMBE R A, STRASDAT H, et al. SLAM++:simultaneous localisation and mapping at the level of objects[C]//Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway:IEEE, 2013:1352-1359. [8] IZADI S, KIM D, HILLIGES O, et al. KinectFusion:real-time 3D reconstruction and interaction using a moving depth camera[C]//Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology. New York:ACM, 2011:559-568. [9] BESCOS B, FÁCIL J M, CIVERA J, et al. DynaSLAM:tracking, mapping, and inpainting in dynamic scenes[J]. IEEE Robotics and Automation Letters, 2018, 3(4):4076-4083. [10] JOHNSON J W. Adapting Mask-RCNN for automatic nucleus segmentation[EB/OL]. (2018-05-01)[2020-11-20]. https://arxiv.org/pdf/1805.00500.pdf. [11] YU C, LIU Z X, LIU X J, et al. DS-SLAM:a semantic visual SLAM towards dynamic environments[C]//Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems. Piscataway:IEEE, 2018:1168-1174. [12] BADRINARAYANAN V, KENDALL A, CIPOLLA R. SegNet:a deep convolutional encoder-decoder architecture for image segmentation[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(12):2481-2495. [13] BOCHKOVSKIY A, WANG C Y, LIAO H Y M. YOLOv4:optimal speed and accuracy of object detection[EB/OL]. (2020-04-23)[2020-11-20]. https://arxiv.org/pdf/2004.10934.pdf. [14] REDMON J, FARHADI A. YOLOv3:an incremental improvement[EB/OL]. (2018-04-08)[2020-11-20]. https://arxiv.org/pdf/1804.02767.pdf. [15] WANG C Y, MARK LIAO H Y, WU Y H, et al. CSPNet:a new backbone that can enhance learning capability of CNN[C]//Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. Piscataway:IEEE, 2020:1571-1580. [16] HE K M, ZHANG X Y, REN S Q, et al. Spatial pyramid pooling in deep convolutional networks for visual recognition[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2015, 37(9):1904-1916. [17] LIU S, QI L, QIN H F, et al. Path aggregation network for instance segmentation[C]//Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway:IEEE, 2018:8759-8768. [18] WANG C Y, BOCHKOVSKIY A, LIAO H Y M. ScaledYOLOv4:scaling cross stage partial network[EB/OL]. (2020-11-16)[2020-11-20]. https://arxiv.org/pdf/2011.08036v1.pdf. [19] STURM J, ENGELHARD N, ENDRES F, et al. A benchmark for the evaluation of RGB-D SLAM systems[C]//Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems. Piscataway:IEEE, 2012:573-580. [20] WHELAN T, SALAS-MORENO R F, GLOCKER B, et al. ElasticFusion:real-time dense SLAM and light source estimation[J]. The International Journal of Robotics Research, 2016, 35(14):1697-1716. [21] RUNZ M, BUFFIER M, AGAPITO L. MaskFusion:real-time recognition, tracking and reconstruction of multiple moving objects[C]//Proceedings of the 2018 IEEE International Symposium on Mixed and Augmented Reality. Piscataway:IEEE, 2018:10-20. |