Stereo camera-based intelligent UGV system for path planning and navigation

被引:0
|
作者
Lee, Jung-Suk [1 ]
Ko, Jung-Hwan [2 ]
Chungb, Dal-Do [3 ]
机构
[1] Inha Tech Coll, Dept Mechatron, 253 Yonghyun Dong, Incheon 402752, South Korea
[2] INCOSYS Co Ltd, Seongnam 463741, Gyunggi, South Korea
[3] Intelligent Hybrid Syst Co Ltd, Juan Venture Ctr, Incheon 402885, South Korea
关键词
stereo target tracking; unmanned ground vehicle; disparity map; pan/tilt system;
D O I
10.1117/12.678788
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In this paper, a new real-time and intelligent mobile robot system for path planning and navigation using stereo camera embedded on the pan/tilt system is proposed. In the proposed system, face area of a moving person is detected from a sequence of the stereo image pairs by using the YCbCr color model and using the disparity map obtained from the left and right images captured by the pan/tilt-controlled stereo camera system and depth information can be detected. And then, the distance between the mobile robot system and the face of the moving person can be calculated from the detected depth information. Accordingly, based-on the analysis of these data, three-dimensional objects can be detected. Finally, by using these detected data, 2-D spatial map for a visually guided robot that can plan paths, navigate surrounding objects and explore an indoor environment is constructed. From some experiments on target tracking with 480 frames of the sequential stereo images, it is analyzed that error ratio between the calculated and measured values of the relative position is found to be very low value of 1.4% on average. Also, the proposed target tracking system has achieved a high speed of 0.04 sec/frame for target detection and 0.06 sec/frame for target tracking.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Stereo camera-based intelligent mobile robot system for path planning and navigation
    Ko, JH
    Jang, JH
    Kim, ES
    [J]. UNMANNED GROUND VEHICLE TECHNOLOGY VII, 2005, 5804 : 677 - 687
  • [2] Tracking Accuracy of a Stereo Camera-Based Augmented Reality Navigation System for Orthognathic Surgery
    Ahn, Jaemyung
    Choi, Hyunseok
    Hong, Jaesung
    Hong, Jongrak
    [J]. JOURNAL OF ORAL AND MAXILLOFACIAL SURGERY, 2019, 77 (05) : 1070.e1 - 1070.e11
  • [3] Intelligent person identification system using stereo camera-based height and stride estimation
    Ko, JH
    Jang, JH
    Kim, ES
    [J]. VISUAL INFORMATION PROCESSING XIV, 2005, 5817 : 85 - 96
  • [4] Stereo Camera-based Computer Vision System of a Robot
    Petrova, Karina
    Filatova, Ekaterina
    [J]. TWELFTH INTERNATIONAL CONFERENCE ON MACHINE VISION (ICMV 2019), 2020, 11433
  • [5] Integrated camera-based navigation
    Hafskjold, BH
    Jalving, B
    Hagen, PE
    Gade, K
    [J]. JOURNAL OF NAVIGATION, 2000, 53 (02): : 237 - 245
  • [6] Wearable RGB Camera-based Navigation System for the Visually Impaired
    Abobeah, Reham
    Hussein, Mohamed
    Abdelwahab, Moataz
    Shoukry, Amin
    [J]. PROCEEDINGS OF THE 13TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISIGRAPP 2018), VOL 5: VISAPP, 2018, : 555 - 562
  • [7] Toward fusion of camera-based egomotion and inertial navigation for a UGV (unmanned ground vehicle) traversing natural terrain
    Haas, G
    Oberle, W
    [J]. UNMANNED GROUND VEHICLE TECHNOLOGY VI, 2004, 5422 : 44 - 54
  • [8] Path planning for UAV/UGV collaborative systems in intelligent manufacturing
    Wang, Qin
    Chen, Hua
    Qiao, Lu
    Tian, Junwei
    Su, Yu
    [J]. IET INTELLIGENT TRANSPORT SYSTEMS, 2020, 14 (11) : 1475 - 1483
  • [9] Automatic Path Planning And Navigation With Stereo Cameras
    Jiang, Shenlu
    Hong, Zhonghua
    Zhang, Yun
    Han, Yanling
    Zhou, Ruyan
    Shen, Bo
    [J]. 2014 THIRD INTERNATIONAL WORKSHOP ON EARTH OBSERVATION AND REMOTE SENSING APPLICATIONS (EORSA 2014), 2014,
  • [10] CAMERA-BASED CLEAR PATH DETECTION
    Wu, Qi
    Zhang, Wende
    Chen, Tsuhan
    Kumar, B. V. K. Vijaya
    [J]. 2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 1874 - 1877