Design and experiments of the binocular visual obstacle perception system for agricultural vehicles

被引:0
|
作者
Wei J. [1 ]
Pan S. [1 ]
Tian G. [2 ]
Gao W. [1 ]
Sun Y. [1 ]
机构
[1] School of Instrument Science and Engineering, Southeast University, Nanjing
[2] College of Engineering, Nanjing Agricultural University, Nanjing
关键词
Agricultural machinery; Depth estimation; Image processing; Obstacle perception; Parallax calculation;
D O I
10.11975/j.issn.1002-6819.2021.09.007
中图分类号
学科分类号
摘要
Machine learning was efficiently incorporated to design a visual perception system for obstacle-free path planning in agricultural vehicles. The present system aims to ensure the safety and reliability of intelligent agricultural vehicles in the process of autonomous navigation. Hardware and software were mainly included in the system. The hardware consisted of visual perception and navigation control module. Since the visual perception task needed real-time image processing, the embedded AI computer Jetson TX2 was taken especially as the core of computing to operate. A deep Convolutional Neural Network (CNN) was used to identify agricultural obstacles. The complex structure and uneven illumination were considered in the agricultural environment, thereby enhancing stability in object detection. The CNN performance of environmental features was much better, compared with the traditional detection using artificially designed features. Moreover, better detection was achieved under continuous learning features in the current task from the large-scale dataset. The improved YOLOv3 was utilized to integrate object detection for the simultaneous output of all information, including category, location, and depth estimation. A binocular camera was used to capture the left and right images, all of which were firstly input into the improved YOLOv3 model for object detection. The output of the improved YOLOv3 model was used for object matching to complete obstacle recognition, where the relationship of obstacles was determined in the left and right images. The location of matching objects was then used to calculate the parallax of the obstacle between left and right images. Finally, the parallax of the obstacle was input into the binocular imaging model for depth estimation. The accuracy of depth estimation was improved, with the increase of model sensitivity to the X-axis of images. The mean error, mean error ratio, and mean square error of depth estimation were greatly improved, compared with the original YOLOv3 and HOG+SVM model. The experimental results showed that the embedded AI computer-processed images in real-time, ensuring the detection accuracy of the improved YOLOv3 model. In object detection, a highly accurate identification was achieved in the agricultural obstacles with an average accuracy rate of 89.54%, and a recall rate of 90.18%. In the first kind of obstacle, the mean error and mean error ratio of the improved YOLOV3 model were 38.92% and 37.23% lower than those of the original one, while 53.44% and 53.14% lower than those of the HOG+SVM model, respectively. In the second kind of obstacle, the mean error and mean error ratio of the improved YOLOV3 model were 26.47% and 26.12% lower than those of the original one, while 41.9% and 41.73% lower than those of the HOG+SVM model, respectively. In the third kind of obstacle, the mean error and mean error ratio of the improved YOLOV3 model were 25.69% and 25.65% lower than those of the original one, while 43.14% and 43.01% lower than those of the HOG+SVM model, respectively. In addition, there was no obvious change in the mean error, mean error ratio, and mean square error of the three models, when changing the distance between obstacle and vehicle. The average error ratio was 4.66% in the depth estimation of obstacles under the dynamic scenario, and the average time was 0.573 s. An electrically controlled hydraulic steering was also used in time for obstacle avoidance during depth warning. The findings can provide an effective basis for environment perception for agricultural vehicles in autonomous navigation. In the following research, the more lightweight YOLOv3-tiny model and the terminal processor Xavier with higher computing power can be selected to conduct the depth estimation, aiming to increase the real-time inference speed of visual perception system in modern agriculture. © 2021, Editorial Department of the Transactions of the Chinese Society of Agricultural Engineering. All right reserved.
引用
收藏
页码:55 / 63
页数:8
相关论文
共 30 条
  • [1] Ji Changying, Zhou Jun, Current situation of navigation technologies for agricultural machinery, Transactions of the Chinese Society for Agricultural Machinery, 45, 9, pp. 44-54, (2014)
  • [2] Zhang Meina, Lv Xiaolan, Tao Jianping, Et al., Design and experiment of automatic guidance control system in agricultural vehicle, Transactions of the Chinese Society for Agricultural Machinery, 47, 7, pp. 42-47, (2016)
  • [3] Zhao Bo, Wang Meng, Mao Enrong, Et al., Recognition and classification for vision navigation application environment of agricultural vehivle, Transactions of the Chinese Society for Agricultural Machinery, 40, 7, pp. 166-170, (2009)
  • [4] Ma Zhaoqing, Yuan Cengren, Real-time navigation and obstacle avoidance based on grids method for fast mobile robot, Robot, 6, pp. 25-29, (1996)
  • [5] Hu Jingtao, Gao Lei, Bai Xiaoping, Et al., Review of research on automatic guidance of agricultural vehicles, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 31, 10, pp. 1-10, (2015)
  • [6] Ji Changying, Shen Ziyao, Gu Baoxing, Et al., Obstacle detection based on point clouds in application of agricultural navigation, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 31, 7, pp. 173-179, (2015)
  • [7] Xue Jinlin, Dong Shuxian, Fan Bowen, Detection of obstacles based on information fusion for autonomous agricultural vehicles, Transactions of the Chinese Society for Agricultural Machinery, 49, S1, pp. 29-34, (2018)
  • [8] Li Le, Zhang Maojun, Xiong Zhihui, Et al., Depth estimation from a single still image of street scene based on content understanding, Robot, 33, 2, pp. 174-180, (2011)
  • [9] Ding Weili, Li Yong, Wang Wenfeng, Et al., Depth estimation of urban road image based on contour understanding, Acta Optica Sinica, 34, 7, pp. 173-179, (2014)
  • [10] Xu Yue, Li Yinghui, Song Huaibo, Et al., Segmentation method of overlapped double apples based on Snake model and corner detectors, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 31, 1, pp. 196-203, (2015)