RETRACTED: Research on multi-task perception network of traffic scene based on feature fusion (Retracted Article)

被引:0
|
作者
Sui, Duo [2 ]
Gao, Peng [2 ]
Fang, Minhang [3 ]
Lian, Jing [1 ,2 ]
Li, Linhui [1 ,2 ]
机构
[1] Dalian Univ Technol, State Key Lab Struct Anal Ind Equipment, Dalian, Liaoning, Peoples R China
[2] Dalian Univ Technol, Sch Automot Engn, Dalian, Liaoning, Peoples R China
[3] Beijing Inst Space Launch Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Traffic scene; multi-task perception; self-attention; feature fusion; intelligent vehicle;
D O I
10.3233/JIFS-235246
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aiming at the problems of low precision and low real-time performance when deploying to embedded platforms in existing multi-task networks, this paper proposes a traffic scene multi-task perception network model (ETS YOLOP) based on feature fusion. Firstly, an Efficient Attention Control Aggregation Network Module (EACAN) is constructed to improve the real-time perception of the model, and the Space Pyramid Pool Fast Convolutional Module (SPPFCSPC) is used at the end of the backbone network to increase the receptive field. Finally, a Multiscale Convolution Transformer Fusion Module (CTFM) is designed in the task branch to better capture global information and rich context information. The experimental results show that compared with the YOLOP model, the ETS YOLOP model has a significant improvement in perception accuracy, 156% in real-time performance, 0.4% increase in mAP on the object detection task, 0.5% increase in mIoU on the drivable area segmentation task, and an 11.4% increase in accuracy on the lane detection task. In order to verify the real-time perception of the model on the embedded platform, the ETS YOLOP model is deployed on the Huawei MDC300F computing platform. Under the condition of the image input size of 640x640, the average frame rate can reach 55FPS, which can realize real-time perception on the embedded platform.
引用
收藏
页码:5753 / 5765
页数:13
相关论文
共 50 条