High-Resolution Fusion Lane Detection Algorithm Based on Model Ensemble

被引:0
|
作者
Yuan X. [1 ]
Jiang C. [1 ]
机构
[1] School of Remote Sensing and Information Engineering, Wuhan University, Wuhan
关键词
convolution neural network; data imbalance; high-resolution fusion; lane detection; model ensemble;
D O I
10.3724/SP.J.1089.2022.19161
中图分类号
学科分类号
摘要
Lane line detection is still a challenging task due to the challenges coming from the complexity of drone aerial images such as complex lane lines, fine-grained feature, class imbalance, etc. Therefore, a lane line detection algorithm based on high-resolution fusion convolution network is proposed. Firstly, the convolution module and up sampling module of full convolution network are improved by using high-resolution fusion structure and bilinear interpolation algorithm. Then, according to the idea of model ensembling, the improved model architecture is used as the foreground-background semantic segmentation model and the multi-category semantic segmentation model, which is used to solve the problem of lane line detection step by step, and the two models are trained by the joint loss function composed of threshold cross entropy loss and Lovasz loss. Finally, the locally region-growth algorithm is used to supplement the details of the detected results. The experimental results show the algorithm achieves 0.548 4 mean intersection over union and 0.993 1 pixel accuracy in the customized drone aerial dataset of 15 types of lane lines, and the prediction speed of 512 × 512 resolution image on NVIDIA Tesla V100 reaches 23.08 frame per second. © 2022 Institute of Computing Technology. All rights reserved.
引用
收藏
页码:1402 / 1410
页数:8
相关论文
共 21 条
  • [1] Neven D, de Brabandere B, Georgoulis S, Et al., Towards end-to-end lane detection: an instance segmentation approach, Proceedings of the IEEE Intelligent Vehicles Symposium, pp. 286-291, (2018)
  • [2] Pan X G, Shi J P, Luo P, Et al., Spatial as deep: spatial CNN for traffic scene understanding, Proceedings of the AAAI Conference on Artificial Intelligence, pp. 7276-7283, (2018)
  • [3] van Gansbeke W, de Brabandere B, Neven D, Et al., End-to-end lane detection through differentiable least-squares fitting, Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, pp. 905-913, (2019)
  • [4] Tabelini L, Berriel R, Paixao T M, Et al., PolyLaneNet: lane estimation via deep polynomial regression, Proceedings of the 5th International Conference on Pattern Recognition, pp. 6150-6156, (2021)
  • [5] Liu R J, Yuan Z J, Liu T, Et al., End-to-end lane shape prediction with transformers, Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3693-3701, (2021)
  • [6] Long J, Shelhamer E, Darrell T., Fully convolutional networks for semantic segmentation, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431-3440, (2015)
  • [7] Sun K, Xiao B, Liu D, Et al., Deep high-resolution representation learning for human pose estimation, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5686-5696, (2019)
  • [8] Berman M, Triki A R, Blaschko M B., The Lovasz-softmax loss: a tractable surrogate for the optimization of the intersection-over-union measure in neural networks, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4413-4421, (2018)
  • [9] He K M, Zhang X Y, Ren S Q, Et al., Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [10] Shi W Z, Caballero J, Huszar F, Et al., Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1874-1883, (2016)