Adapt-Net: A Unified Object Detection Framework for Mobile Augmented Reality

被引:0
|
作者
Zeng, Xiangyun [1 ]
Tan, Siok Yee [1 ]
Nasrudin, Mohammad Faidzul [1 ]
机构
[1] Univ Kebangsaan Malaysia, Fac Informat Sci & Technol, Ctr Artificial Intelligence Technol, Bangi 43600, Malaysia
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Adaptation models; Object detection; Computational modeling; Accuracy; Data models; Feature extraction; Training; Contrastive learning; Deep learning; Mutual information; Knowledge management; Augmented reality; deep mutual learning; masked generative knowledge distillation; mobile augmented reality; object detection; unsupervised domain adaptation;
D O I
10.1109/ACCESS.2024.3447043
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection is a crucial task in mobile augmented reality (MAR), where achieving both speed and accuracy with limited computational resources is essential. However, applying object detection models to new domains or reducing the model size tends to lower their performance. To address this problem, this research introduced a unified object detection framework called Adapt-Net. This framework incorporates contrastive learning techniques for unsupervised domain adaptation, a teacher-student structure generative compressed model with masking, and deep mutual learning of student models, all built upon the YOLOv8 architecture. Adapt-Net's key novelty lies in its unified framework that combines three models: two student models and one teacher model. Each model comprises a feature-extracting backbone and an adapter network. The student models backbone are trained using deep mutual learning and contrastive learning loss to ensure domain-invariant feature generation. Unsupervised domain adaptation and masked generative knowledge distillation modules facilitate knowledge transfer from the teacher to the student models, enhancing their ability to generalize to unfamiliar objects. The use of masked generative knowledge distillation, which guides the student models to reconstruct the teacher's features from a masked input in a generative manner, rather than merely imitating the output. This generative approach improves the student models' representation capabilities. Adapt-Net enables the student models to not only learn domain-invariant features but also enhance their generalization capabilities to new objects. Extensive experiments conducted on benchmark datasets demonstrate that our proposed approach surpasses state-of-the-art object detection methods by 6.8 mAP score in terms of detection accuracy on the Microsoft COCO dataset. Notably, the model size remains a compact 3.2M, enabling fast inference speeds, lower computational resource consumption, and enhanced resilience to domain variations. Adapt-Net represents a promising and efficient approach to object detection that combines accuracy with efficiency.
引用
收藏
页码:120788 / 120803
页数:16
相关论文
共 50 条
  • [1] Object Detection in the Context of Mobile Augmented Reality
    Li, Xiang
    Tian, Yuan
    Zhang, Fuyao
    Quan, Shuxue
    Xu, Yi
    2020 IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY (ISMAR 2020), 2020, : 156 - 163
  • [2] A Review of Lightweight Object Detection Algorithms for Mobile Augmented Reality
    Nafea, Mohammed Mansoor
    Tan, Siok Yee
    Jubair, Mohammed Ahmed
    Abd, Mustafa Tareq
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (11) : 536 - 546
  • [3] A framework for outdoor mobile augmented reality
    FSCSHD, CoESTAR, UNIMAS, Kota Samarahan, Sarawak, Malaysia
    Int. J. Comput. Sci. Issues, 2 (419-423):
  • [4] Mobile Augmented Reality Framework - MIRAR
    Rodrigues, Joao M. F.
    Veiga, Ricardo J. M.
    Bajireanu, Roman
    Lam, Roberto
    Pereira, Joao A. R.
    Sardo, Joao D. P.
    Cardoso, Pedro J. S.
    Bica, Paulo
    UNIVERSAL ACCESS IN HUMAN-COMPUTER INTERACTION: VIRTUAL, AUGMENTED, AND INTELLIGENT ENVIRONMENTS, 2018, 10908 : 102 - 121
  • [5] An Improved YOLOv3 Object Detection Network for Mobile Augmented Reality
    Wang, Quanyu
    Wang, Zhi
    Li, Bei
    Wei, Dejian
    2021 IEEE 7TH INTERNATIONAL CONFERENCE ON VIRTUAL REALITY (ICVR 2021), 2021, : 332 - 339
  • [6] Edge Assisted Real-time Object Detection for Mobile Augmented Reality
    Liu, Luyang
    Li, Hongyu
    Gruteser, Marco
    MOBICOM'19: PROCEEDINGS OF THE 25TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, 2019,
  • [7] Frugal Following: Power Thrifty Object Detection and Tracking for Mobile Augmented Reality
    Apicharttrisorn, Kittipat
    Ran, Xukan
    Chen, Jiasi
    Krishnamurthy, Srikanth, V
    Roy-Chowdhury, Amit K.
    PROCEEDINGS OF THE 17TH CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS (SENSYS '19), 2019, : 96 - 109
  • [8] A framework for context immersion in mobile augmented reality
    Kim, Mi Jeong
    AUTOMATION IN CONSTRUCTION, 2013, 33 : 79 - 85
  • [9] MARS: Mobile Augmented Reality Script Framework
    Constancio, Diogo
    Casanova, Daniel
    Fernandes, Antonio Ramires
    2017 24 ENCONTRO PORTUGUES DE COMPUTACAO GRAFICA E INTERACAO (EPCGI), 2017,
  • [10] Mobile Augmented Reality Tourism Application Framework
    Abd Rashid, Rashidi
    Mohamed, Halina
    Hussin, Ab Razak Che
    RECENT TRENDS IN INFORMATION AND COMMUNICATION TECHNOLOGY, 2018, 5 : 108 - 115