Adapt-Net: A Unified Object Detection Framework for Mobile Augmented Reality

被引:0
|
作者
Zeng, Xiangyun [1 ]
Tan, Siok Yee [1 ]
Nasrudin, Mohammad Faidzul [1 ]
机构
[1] Univ Kebangsaan Malaysia, Fac Informat Sci & Technol, Ctr Artificial Intelligence Technol, Bangi 43600, Malaysia
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Adaptation models; Object detection; Computational modeling; Accuracy; Data models; Feature extraction; Training; Contrastive learning; Deep learning; Mutual information; Knowledge management; Augmented reality; deep mutual learning; masked generative knowledge distillation; mobile augmented reality; object detection; unsupervised domain adaptation;
D O I
10.1109/ACCESS.2024.3447043
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection is a crucial task in mobile augmented reality (MAR), where achieving both speed and accuracy with limited computational resources is essential. However, applying object detection models to new domains or reducing the model size tends to lower their performance. To address this problem, this research introduced a unified object detection framework called Adapt-Net. This framework incorporates contrastive learning techniques for unsupervised domain adaptation, a teacher-student structure generative compressed model with masking, and deep mutual learning of student models, all built upon the YOLOv8 architecture. Adapt-Net's key novelty lies in its unified framework that combines three models: two student models and one teacher model. Each model comprises a feature-extracting backbone and an adapter network. The student models backbone are trained using deep mutual learning and contrastive learning loss to ensure domain-invariant feature generation. Unsupervised domain adaptation and masked generative knowledge distillation modules facilitate knowledge transfer from the teacher to the student models, enhancing their ability to generalize to unfamiliar objects. The use of masked generative knowledge distillation, which guides the student models to reconstruct the teacher's features from a masked input in a generative manner, rather than merely imitating the output. This generative approach improves the student models' representation capabilities. Adapt-Net enables the student models to not only learn domain-invariant features but also enhance their generalization capabilities to new objects. Extensive experiments conducted on benchmark datasets demonstrate that our proposed approach surpasses state-of-the-art object detection methods by 6.8 mAP score in terms of detection accuracy on the Microsoft COCO dataset. Notably, the model size remains a compact 3.2M, enabling fast inference speeds, lower computational resource consumption, and enhanced resilience to domain variations. Adapt-Net represents a promising and efficient approach to object detection that combines accuracy with efficiency.
引用
收藏
页码:120788 / 120803
页数:16
相关论文
共 50 条
  • [21] Automatic text detection for mobile augmented reality translation
    Petter, Marc
    Fragoso, Victor
    Turk, Matthew
    Baur, Charles
    2011 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCV WORKSHOPS), 2011,
  • [22] Adaptive Multimodal Interaction in Mobile Augmented Reality: A Conceptual Framework
    Abidin, Rimaniza Zainal
    Arshad, Haslina
    Shukri, Saidatul A'isyah Ahmad
    2ND INTERNATIONAL CONFERENCE ON APPLIED SCIENCE AND TECHNOLOGY 2017 (ICAST'17), 2017, 1891
  • [23] MIRAR: Mobile Image Recognition Based Augmented Reality Framework
    Pereira, Joao A. R.
    Veiga, Ricardo J. M.
    de Freitas, Marco A. G.
    Sardo, J. D. P.
    Cardoso, Pedro J. S.
    Rodrigues, Joao M. F.
    INCREASE, 2018, : 321 - 337
  • [24] CloudAR: A Cloud-based Framework for Mobile Augmented Reality
    Zhang, Wenxiao
    Lin, Sikun
    Bijarbooneh, Farshid Hassani
    Cheng, Hao Fei
    Hui, Pan
    PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, : 194 - 200
  • [25] PrivacyManager: An Access Control Framework for Mobile Augmented Reality Applications
    Lehman, Sarah M.
    Tan, Chiu C.
    2017 IEEE CONFERENCE ON COMMUNICATIONS AND NETWORK SECURITY (CNS), 2017, : 28 - 36
  • [26] Object detection networks and augmented reality for cellular detection in fluorescence microscopy
    Waithe, Dominic
    Brown, Jill M.
    Reglinski, Katharina
    Diez-Sevilla, Isabel
    Roberts, David
    Eggeling, Christian
    JOURNAL OF CELL BIOLOGY, 2020, 219 (10):
  • [27] An Efficient Object Augmentation Scheme for Supporting Pervasiveness in a Mobile Augmented Reality
    Jang, Sung-Bong
    Ko, Young-Woong
    JOURNAL OF INFORMATION PROCESSING SYSTEMS, 2020, 16 (05): : 1214 - 1222
  • [28] Mobile Augmented Reality: Fast, Precise, and Smooth Planar Object Tracking
    Matveichev, Dmitrii
    Lin, Daw-Tung
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 5406 - 5412
  • [29] Object Management Based on Metadata Registry for Intelligent Mobile Augmented Reality
    Jang, Sung-Bong
    2019 1ST INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION (ICAIIC 2019), 2019, : 572 - 574
  • [30] Marker Classification Method for Hierarchical Object Navigation in Mobile Augmented Reality
    Park, Gyeong-Mi
    Han, PhyuPhyu
    Kim, Youngbong
    MULTIMEDIA, COMPUTER GRAPHICS AND BROADCASTING, PT I, 2011, 262 : 179 - 184