Fast Armored Target Detection Based on Lightweight Network

被引:0
|
作者
Sun H. [1 ,2 ]
Chang T. [1 ]
Zhang L. [1 ]
Yang G. [1 ]
Han B. [1 ]
Li Y. [3 ]
机构
[1] Department of Weapon and Control, Army Academy of Armored Forces, Beijing
[2] 78123 Troop of the PLA, Chengdu
[3] The 2rd District, Army Base of Test and Training, Weinan
关键词
Armored target; Lightweight convolutional neural network; One-stage detector; Target detection;
D O I
10.3724/SP.J.1089.2019.17467
中图分类号
学科分类号
摘要
Focused on the detection task of armored target in battlefield environment, a fast detection method based on lightweight convolutional neural network is proposed in this paper. Firstly, based on the lightweight backbone network (MobileNet), a multi-scale single-stage detection framework is developed. Secondly, considering the size distribution of armored target, higher resolution feature maps are selected and a new designed Resblock is added to each detection unit to enhance the detection performance for small targets. At last, focal-loss function is introduced to replace the traditional cross entropy loss function, which effectively overcomes the extreme imbalance of the distribution of the positive and negative samples in training processes. A special detection dataset for armored target is constructed, based on which the comparable experiments with state-of-art detection methods are conducted. Experimental results show that the proposed method achieves good performance in detection accuracy, model size and operation speed, and is especially suitable for small mobile reconnaissance platforms such as UAVs (unmanned aerial vehicle). © 2019, Beijing China Science Journal Publishing Co. Ltd. All right reserved.
引用
收藏
页码:1110 / 1121
页数:11
相关论文
共 23 条
  • [11] Liu W., Anguelov D., Erhan D., Et al., SSD: single shot multiBox detector, Proceedings of the European Conference on Computer Vision, pp. 21-37, (2016)
  • [12] Redmon J., Farhadi A., YOLO9000: better, faster, stronger, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6517-6525, (2017)
  • [13] Fu C.Y., Liu W., Ranga A., Et al., DSSD: deconvolutional single shot detector
  • [14] Everingham M., Van Gool L., Williams C.K.I., Et al., The pascal visual object classes (VOC) challenge, International Journal of Computer Vision, 88, 2, pp. 303-338, (2010)
  • [15] Geiger A., Lenz P., Urtasun R., Are we ready for autonomous driving? The Kitti vision benchmark suite, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354-3361, (2012)
  • [16] Simonyan K., Zisserman A., Very deep convolutional networks for large-scale image recognition
  • [17] He K.M., Zhang X.Y., Ren S.Q., Et al., Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [18] Lin T.Y., Goyal P., Girshick R., Et al., Focal loss for dense object detection, Proceedings of the IEEE International Conference on Computer Vision, pp. 2999-3007, (2017)
  • [19] Iandola F.N., Han S., Moskewicz M.W., Et al., Squeezenet: Alexnet-level accuracy with 50x fewer parameters and< 0.5 MB model size
  • [20] Howard A.G., Zhu M.L., Chen B., Et al., MobileNets: efficient convolutional neural networks for mobile vision applications