1 |
CHEN S Y, LI Y F, KWOK N M. Active vision in robotic systems: a survey of recent developments[J]. The International Journal of Robotics Research, 2011, 30(11): 1343-1377. 10.1177/0278364911410755
|
2 |
SHI H B, SUN G, WANG Y P, et al. Adaptive image-based visual servoing with temporary loss of the visual signal[J]. IEEE Transactions on Industrial Informatics, 2019, 15(4): 1956-1965. 10.1109/tii.2018.2865004
|
3 |
CAO Z Q, CHEN X C, YU Y Y, et al. Image dynamics-based visual servoing for quadrotors tracking a target with a nonlinear trajectory observer[J]. IEEE Transactions on System, Man, and Cybernetics: Systems, 2020, 50(1): 376-384. 10.1109/tsmc.2017.2720173
|
4 |
SALEHI I, ROTITHOR G, SALTUS R, et al. Constrained image-based visual servoing using barrier functions [C]// Proceedings of the 2021 IEEE International Conference on Robotics and Automation. Piscataway: IEEE, 2021: 14254-14260. 10.1109/icra48506.2021.9560983
|
5 |
LIANG X W, WANG H S, LIU Y H, et al. Formation control of nonholonomic mobile robots without position and velocity measurements[J]. IEEE Transactions on Robotics, 2018, 34(2): 434-446. 10.1109/tro.2017.2776304
|
6 |
WANG H S, YANG B H, LIU Y T, et al. Visual servoing of soft robot manipulator in constrained environments with an adaptive controller[J]. IEEE/ASME Transactions on Mechatronics, 2017, 22(1): 41-50. 10.1109/tmech.2016.2613410
|
7 |
ZHENG D L, WANG H S, WANG J C, et al. Image-based visual servoing of a quadrotor using virtual camera approach[J]. IEEE/ASME Transactions on Mechatronics, 2017, 22(2): 972-982. 10.1109/tmech.2016.2639531
|
8 |
AGRAVANTE D J, CHAUMETTE F. Active vision for pose estimation applied to singularity avoidance in visual servoing [C]// Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems. Piscataway: IEEE, 2017: 2947-2952. 10.1109/iros.2017.8206129
|
9 |
ZHANG K X, CHEN J, LI Y, et al. Visual tracking and depth estimation of mobile robots without desired velocity information[J]. IEEE Transactions on Cybernetics, 2020, 50(1): 361-373. 10.1109/tcyb.2018.2869623
|
10 |
LI J N, XIE H, LOW K H, et al. Image-based visual servoing of rotorcrafts to planar visual targets of arbitrary orientation[J]. IEEE Robotics and Automation Letters, 2021, 6(4): 7861-7868. 10.1109/lra.2021.3101878
|
11 |
ESPIAU B, CHAUMETTE F, RIVES P. A new approach to visual servoing in robotics[J]. IEEE Transactions on Robotics and Automation, 1992, 8(3): 313-326. 10.1109/70.143350
|
12 |
徐德,卢金燕.直线特征的交互矩阵求取[J].自动化学报, 2015, 41(10): 1762-1771. 10.16383/j.aas.2015.c150097
|
|
XU D, LU J Y. Determination for interactive matrix of line feature[J]. Acta Automatica Sinica, 2015, 41(10): 1762-1771. 10.16383/j.aas.2015.c150097
|
13 |
XU C, ZHANG L L, CHENG L, et al. Pose estimation from line correspondences: a complete analysis and a series of solutions[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(6): 1209-1222. 10.1109/tpami.2016.2582162
|
14 |
ZHANG G X, LEE J H, LIM J, et al. Building a 3-D line-based map using stereo SLAM[J]. IEEE Transaction on Robotics, 2015, 31(6): 1364-1377. 10.1109/tro.2015.2489498
|
15 |
ZHOU H Z, ZOU D P, PEI L, et al. StructSLAM: visual SLAM with building structure lines[J]. IEEE Transactions on Vehicular Technology, 2015, 64(4): 1364-1375. 10.1109/tvt.2015.2388780
|
16 |
MATEUS A, TAHRI O, MIRALDO P. Active estimation of 3D lines in spherical coordinates [C]// Proceedings of the 2019 American Control Conference. Piscataway: IEEE, 2019: 3950-3955. 10.23919/acc.2019.8814938
|
17 |
DE FARIAS C, ADJIGBLE M, TAMADAZTE B, et al. Dual quaternion-based visual servoing for grasping moving objects [C]// Proceedings of the IEEE 17th International Conference on Automation Science and Engineering. Piscataway: IEEE, 2021: 151-158. 10.1109/case49439.2021.9551631
|
18 |
FU Q, YU H S, LAI L H, et al. A robust RGB-D SLAM system with points and lines for low texture indoor environments[J]. IEEE Sensors Journal, 2019, 19(21): 9908-9920. 10.1109/jsen.2019.2927405
|
19 |
PUMAROLA A, VAKHITOV A, AGUDO A, et al. PL-SLAM: real-time monocular visual SLAM with points and lines [C]// Proceedings of the 2017 IEEE International Conference on Robotics and Automation. Piscataway: IEEE, 2017: 4503-4508. 10.1109/icra.2017.7989522
|
20 |
HE Y J, ZHAO J, GUO Y, et al. PL-VIO: tightly-coupled monocular visual-inertial odometry using point and line features[J]. Sensors, 2018, 18(4): No.1159. 10.3390/s18041159
|
21 |
WANG R Z, DI K C, WAN W H, et al. Improved point-line feature based visual SLAM method for indoor scenes[J]. Sensors, 2018, 18(10): No.3559. 10.3390/s18103559
|
22 |
HUANG S S, MA Z Y, MU T J, et al. Lidar-monocular visual odometry using point and line features [C]// Proceedings of the 2020 IEEE International Conference on Robotics and Automation. Piscataway: IEEE, 2020: 1091-1097. 10.1109/icra40945.2020.9196613
|
23 |
WANG R H, ZHANG X B, FANG Y C, et al. Virtual-goal-guided RRT for visual servoing of mobile robots with FOV constraint[J]. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2022, 52(4): 2073-2083. 10.1109/tsmc.2020.3044347
|