Real-time semantic segmentation based on improved BiSeNet

被引:2
|
作者
Ren F. [1 ,2 ]
Yang L. [1 ,2 ]
Zhou H. [1 ,2 ]
Zhang S. [1 ,2 ]
He X. [3 ]
Xu W. [4 ]
机构
[1] Tianjin Key Laboratory for Advanced Mechatronic System Design and Intelligent Control, School of Mechanical Engineering, Tianjin University of Technology, Tianjin
[2] National Demonstration Center for Experimental Mechanical and Electrical Engineering Education, Tianjin University of Technology, Tianjin
[3] Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun
[4] Transcend Communication Technology Tianjin Co., Ltd, Tianjin
关键词
attention mechanism; deep learning; real time; semantic segmentation;
D O I
10.37188/OPE.20233108.1217
中图分类号
学科分类号
摘要
To improve the performance of image semantic segmentation on accuracy and efficiency for practical applications, in this study, we propose a real-time semantic segmentation algorithm based on improved BiSeNet. First, the redundancy of certain channels and parameters of BiSeNet is eliminated by sharing the heads of dual branches, and the affluent shallow features are effectively extracted at the same time. Subsequently, the shared layers are divided into dual branches, namely, the detail branch and the semantic branch, which are used to extract detailed spatial information and contextual semantic information, respectively. Furthermore, both the channel attention mechanism and spatial attention mechanism are introduced into the tail of the semantic branch to enhance the feature representation; thus the BiSeNet is optimized by using dual attention mechanisms to extract contextual semantic features more effectively. Finally, the features of the detail branch and semantic branch are fused and up-sampled to the resolution of the input image to obtain semantic segmentation. Our proposed algorithm achieves 77. 2% mIoU on accuracy with real-time performance of 95. 3 FPS on Cityscapes dataset and 73. 8% mIoU on accuracy with real-time performance of 179. 1 FPS on CamVid dataset. The experiments demonstrate that our proposed semantic segmentation algorithm achieves a good trade-off between accuracy and efficiency. Furthermore, the performance of semantic segmentation is significantly improved compared with BiSeNet and other existing algorithms. © 2023 Chinese Academy of Sciences. All rights reserved.
引用
下载
收藏
页码:1217 / 1227
页数:10
相关论文
共 36 条
  • [1] WANG ZH Y, NI X Y, SHANG ZH D., Autonomous driving semantic segmentation with convolution neural networks, Opt. Precision Eng, 27, 11, pp. 2429-2438, (2019)
  • [2] RONNEBERGER O, FISCHER P, BROX T., UNet: Convolutional Networks For Biomedical Image Segmentation, Lecture Notes in Computer Science, pp. 234-241, (2015)
  • [3] CHEN J, LU Y, YU Q, Et al., TransUNet: Transformers Make Strong Encoders For Medical Image Segmentation [EB/OL], (2021)
  • [4] WANG Y N, WANG T F, TIAN Y Z, Et al., Improved local convexity algorithm of segmentation for 3D point cloud, Chinese Optics, 10, 3, pp. 348-354, (2017)
  • [5] REN F L, ZHOU H B, YANG L, Et al., ADPNet: Attention based dual path network for lane detection, Journal of Visual Communication and Image Representation, 87, (2022)
  • [6] LONG J, SHELHAMER E, DARRELL T., Fully Convolutional Networks for Semantic Segmentation [C], 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3431-3440, (2015)
  • [7] CHEN L C, PAPANDREOU G, KOKKINOS I, Et al., DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, IEEE Transactions on Pattern Analysis and Machine Intelligence, 40, 4, pp. 834-848, (2018)
  • [8] CHEN L C, PAPANDREOU G, SCHROFF F, Et al., Rethinking Atrous Convolution for Semantic Image Segmentation [EB/OL], (2017)
  • [9] CHEN L C, ZHU Y K, PAPANDREOU G, Et al., Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation, Computer Vision-ECCV 2018, pp. 833-851, (2018)
  • [10] REN F L, HE X, WEI ZH H, Et al., Semantic segmentation based on DeepLabV3+ and superpixel optimization, Opt. Precision Eng, 27, 12, pp. 2722-2729, (2019)