Learning Balance Feature for Object Detection

被引:0
|
作者
Zhang, Zhiqiang [1 ,2 ]
Qiu, Xin [1 ]
Li, Yongzhou [1 ]
机构
[1] Chinese Acad Sci, Inst Microelect, 3 Beitucheng West Rd, Beijing 100029, Peoples R China
[2] Univ Chinese Acad Sci, 19 A Yuquan Rd, Beijing 100049, Peoples R China
关键词
object detection; Feature Pyramid Network; feature transformer; feature balance; FPN;
D O I
10.3390/electronics11172765
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of studying scale variation, the Feature Pyramid Network (FPN) replaces the image pyramid and has become one of the most popular object detection methods for detecting multi-scale objects. State-of-the-art methods have FPN inserted into a pipeline between the backbone and the detection head to enable shallow features with more semantic information. However, FPN is insufficient for object detection on various scales, especially for small-scale object detection. One of the reasons is that the features are extracted at different network depths, which introduces gaps between features. That is, as the network becomes deeper and deeper, the high-level features have more semantics but less content description. This paper proposes a new method that includes a multi-scale receptive fields extraction module, a feature constructor module, and an attention module to improve the detection efficiency of FPN for objects of various scales and to bridge the gap in content description and semantics between different layers. Together, these three modules make the detector capable of selecting the most suitable feature for objects. Especially for the attention module, this paper chooses to use a parallel structure to simultaneously extract channel and spatial attention from the same features. When we use Adopting Adaptive Training Sample Selection (ATSS) and FreeAnchor as the baseline and ResNet50 as the backbone, the experimental results on the MS COCO dataset show that our algorithm can enhance the mean average precision (mAP) by 3.7% and 2.4% compared to FPN, respectively.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Feature Enhancement SSD for Object Detection
    Tan H.
    Li S.
    Liu B.
    Liu X.
    [J]. Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2019, 31 (04): : 573 - 579
  • [42] Object Detection In Quantized Feature Space
    Bulla, Christopher
    Luthra, Bhomik
    Qian, Ningqing
    [J]. 2014 IEEE FOURTH INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS BERLIN (ICCE-BERLIN), 2014, : 391 - 394
  • [43] On Semantic Object Detection with Salient Feature
    Li, Zhidong
    Chen, Jing
    [J]. ADVANCES IN VISUAL COMPUTING, PT II, PROCEEDINGS, 2008, 5359 : 782 - 791
  • [44] Foreground Feature Enhancement for Object Detection
    Jiang, Shenwang
    Xu, Tingfa
    Li, Jianan
    Shen, Ziyi
    Guo, Jie
    [J]. IEEE ACCESS, 2019, 7 : 49223 - 49231
  • [45] Feature Selective Networks for Object Detection
    Zhai, Yao
    Fu, Jingjing
    Lu, Yan
    Li, Houqiang
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4139 - 4147
  • [46] Centralized Feature Pyramid for Object Detection
    Quan, Yu
    Zhang, Dong
    Zhang, Liyan
    Tang, Jinhui
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 4341 - 4354
  • [47] Robust feature design for object detection
    Hu, Woong
    Koo, Min-Su
    Nam, Jae-Hyun
    Kim, Byung-Gyu
    Kim, Sung-Ki
    [J]. Lecture Notes in Electrical Engineering, 2015, 373 : 117 - 123
  • [48] Masked Feature Compression for Object Detection
    Dai, Chengjie
    Song, Tiantian
    Jin, Yuxuan
    Ren, Yixiang
    Yang, Bowei
    Song, Guanghua
    [J]. MATHEMATICS, 2024, 12 (12)
  • [49] Adaptive multiscale feature for object detection
    Yu, Xiaoyong
    Wu, Siyuan
    Lu, Xiaoqiang
    Gao, Guilong
    [J]. NEUROCOMPUTING, 2021, 449 : 146 - 158
  • [50] Feature Pyramid Networks for Object Detection
    Lin, Tsung-Yi
    Dollar, Piotr
    Girshick, Ross
    He, Kaiming
    Hariharan, Bharath
    Belongie, Serge
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 936 - 944