Anchor-free lightweight infrared object detection method (Invited)

被引:0
|
作者
Gao F. [1 ,2 ]
Yang X. [2 ]
Lu R. [2 ]
Wang S. [2 ]
Gao J. [2 ]
Xia H. [2 ]
机构
[1] Beijing Huahang Radio Measurement Institute, Beijing
[2] Missile Engineering Institute, Rocket Force University of Engineering, Xi'an
关键词
Asymmetric convolution; Infrared target; Lightweight; Neural network; Object detection;
D O I
10.3788/IRLA20220193
中图分类号
学科分类号
摘要
According to the characteristics of infrared targets, an anchor-free lightweight infrared target detection method was proposed, which improved the detection ability of embedded platform. For the platform with limited computing resources, a new lightweight convolution structure was proposed. Asymmetric convolution was introduced to enhance the feature expression ability of standard convolution, reducing the amount of parameters and computation effectively. A lightweight feature extraction unit was constructed by designing parallel multi-feature path, which generated rich features through channel concatation, then combining with attention module and channel shuffle. SkipBranch was added to promote the transmission of shallow information to the high level and further enrich the characteristics of the high level. Experiments on FLIR dataset showed that the accuracy of the designed lightweight network structure was 81.7%, which exceeded YOLOv4-tiny. However, the model parameters and calculation amount were reduced by 75.0% and 71.1% respectively, and the reasoning time was compressed by 91.3%, which could meet the real-time detection requirements of infrared object on embedded platform. Copyright ©2022 Infrared and Laser Engineering. All rights reserved.
引用
收藏
相关论文
共 22 条
  • [1] Howard A G, Zhu M, Chen B, Et al., MobileNets: Efficient convolutional neural networks for mobile vision applications [J], (2017)
  • [2] Sandler M, Howard A, Zhu M, Et al., MobileNetV2: Inverted residuals and linear bottlenecks, IEEE/CVF Conference on Computer Vision & Pattern Recognition, pp. 4510-4520, (2018)
  • [3] Howard A, Sandler M, Chen B, Et al., Searching for MobileNetV3, IEEE/CVF International Conference on Computer Vision, pp. 1314-1324, (2019)
  • [4] Hu Jie, Shen Li, Sun Gang, Et al., Squeeze-and-excitation networks, IEEE/CVF Conference on Computer Vision & Pattern Recognition, pp. 7132-7141, (2018)
  • [5] Zhang X, Zhou X, Lin M, Et al., ShuffleNet: An extremely efficient convolutional neural network for mobile devices, CVF Conference on Computer Vision & Pattern Recognition, pp. 6848-6856, (2018)
  • [6] Ma N, Zhang X, Zheng H T, Et al., ShuffleNetV2: Practical guidelines for efficient CNN architecture design, European Conference on Computer Vision, 11218, pp. 122-138, (2018)
  • [7] Iandola F N, Han S, Moskewicz M W, Et al., SqueezeNet: AlexNet-level accuracy with 50 x fewer parameters and <0.5 MB model size [J], (2016)
  • [8] Han K, Wang Y, Tian Q, Et al., GhostNet: More features from cheap operations, CVF Conference on Computer Vision & Pattern Recognition, pp. 1577-1586, (2020)
  • [9] Tan M X, Le Q V., EfficientNet: Rethinking model scaling for convolutional neural networks [J], (2019)
  • [10] Tan M X, Le Q V., EfficientNetV2: Smaller models and faster training [J], (2021)