Plug and Play Active Learning for Object Detection

被引:5
|
作者
Yang, Chenhongyi [1 ]
Huang, Lichao [2 ]
Crowley, Elliot J. [1 ]
机构
[1] Univ Edinburgh, Sch Engn, Edinburgh, Midlothian, Scotland
[2] Horizon Robot, Beijing, Peoples R China
关键词
D O I
10.1109/CVPR52733.2024.01684
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Annotating datasets for object detection is an expensive and time-consuming endeavor. To minimize this burden, active learning (AL) techniques are employed to select the most informative samples for annotation within a constrained "annotation budget". Traditional AL strategies typically rely on model uncertainty or sample diversity for query sampling, while more advanced methods have focused on developing AL-specific object detector architectures to enhance performance. However, these specialized approaches are not readily adaptable to different object detectors due to the significant engineering effort required for integration. To overcome this challenge, we introduce Plug and Play Active Learning (PPAL), a simple and effective AL strategy for object detection. PPAL is a two-stage method comprising uncertainty-based and diversity-based sampling phases. In the first stage, our Difficulty Calibrated Uncertainty Sampling leverage a category-wise difficulty coefficient that combines both classification and localisation difficulties to re-weight instance uncertainties, from which we sample a candidate pool for the subsequent diversity-based sampling. In the second stage, we propose Category Conditioned Matching Similarity to better compute the similarities of multi-instance images as ensembles of their instance similarities, which is used by the k- Means++ algorithm to sample the final AL queries. PPAL makes no change to model architectures or detector training pipelines; hence it can be easily generalized to different object detectors. We benchmark PPAL on the MS-COCO and Pascal VOC datasets using different detector architectures and show that our method outperforms prior work by a large margin. Code is available at https://github.com/ChenhongyiYang/PPAL
引用
收藏
页码:17784 / 17793
页数:10
相关论文
共 50 条
  • [31] Active Structured Learning for High-Speed Object Detection
    Lampert, Christoph H.
    Peters, Jan
    PATTERN RECOGNITION, PROCEEDINGS, 2009, 5748 : 221 - 231
  • [32] Self-Supervised Reinforcement Learning for Active Object Detection
    Fang, Fen
    Liang, Wenyu
    Wu, Yan
    Xu, Qianli
    Lim, Joo-Hwee
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04): : 10224 - 10231
  • [33] Active Learning for Deep Object Detection via Probabilistic Modeling
    Choi, Jiwoong
    Elezi, Ismail
    Lee, Hyuk-Jae
    Farabet, Clement
    Alvarez, Jose M.
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10244 - 10253
  • [34] PLUG-AND-PLAY PROGRAMMING - AN OBJECT-ORIENTED CONSTRUCTION KIT
    COTE, RGA
    BYTE, 1994, 19 (06): : 42 - 42
  • [35] FAIME: an object-oriented methodology for application plug-and-play
    Univ of North Carolina, Charlotte, United States
    JOOP J Object Oriented Program, 5 (8pp):
  • [36] Plug-in Approach to Active Learning
    Minsker, Stanislav
    JOURNAL OF MACHINE LEARNING RESEARCH, 2012, 13 : 67 - 90
  • [37] Active Learning as a Tool for Optimizing "Plug-n-Play" Electrochemical Atom Transfer Radical Polymerization
    Zhao, Boyu
    Cheng, Jiahao
    Gao, Junlong
    Haddleton, David. M.
    Wilson, Paul
    MACROMOLECULAR CHEMISTRY AND PHYSICS, 2023, 224 (12)
  • [38] Stochastic Fault Detection in a Plug-and-Play Scenario
    Poem, Francesca
    Riverso, Stefano
    Ferrari-Trecate, Giancarlo
    Parisini, Thomas
    2015 54TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2015, : 3137 - 3142
  • [39] Plug-and-Play Joint Image Deblurring and Detection
    Marrs, Corey
    Kathariya, Birendra
    Li, Zhu
    York, George
    2023 IEEE 25TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, MMSP, 2023,
  • [40] Active object detection
    de Croon, G.
    VISAPP 2007: Proceedings of the Second International Conference on Computer Vision Theory and Applications, Volume IU/MTSV, 2007, : 97 - 103