Adapt-Net: A Unified Object Detection Framework for Mobile Augmented Reality

被引:0
|
作者
Zeng, Xiangyun [1 ]
Tan, Siok Yee [1 ]
Nasrudin, Mohammad Faidzul [1 ]
机构
[1] Univ Kebangsaan Malaysia, Fac Informat Sci & Technol, Ctr Artificial Intelligence Technol, Bangi 43600, Malaysia
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Adaptation models; Object detection; Computational modeling; Accuracy; Data models; Feature extraction; Training; Contrastive learning; Deep learning; Mutual information; Knowledge management; Augmented reality; deep mutual learning; masked generative knowledge distillation; mobile augmented reality; object detection; unsupervised domain adaptation;
D O I
10.1109/ACCESS.2024.3447043
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection is a crucial task in mobile augmented reality (MAR), where achieving both speed and accuracy with limited computational resources is essential. However, applying object detection models to new domains or reducing the model size tends to lower their performance. To address this problem, this research introduced a unified object detection framework called Adapt-Net. This framework incorporates contrastive learning techniques for unsupervised domain adaptation, a teacher-student structure generative compressed model with masking, and deep mutual learning of student models, all built upon the YOLOv8 architecture. Adapt-Net's key novelty lies in its unified framework that combines three models: two student models and one teacher model. Each model comprises a feature-extracting backbone and an adapter network. The student models backbone are trained using deep mutual learning and contrastive learning loss to ensure domain-invariant feature generation. Unsupervised domain adaptation and masked generative knowledge distillation modules facilitate knowledge transfer from the teacher to the student models, enhancing their ability to generalize to unfamiliar objects. The use of masked generative knowledge distillation, which guides the student models to reconstruct the teacher's features from a masked input in a generative manner, rather than merely imitating the output. This generative approach improves the student models' representation capabilities. Adapt-Net enables the student models to not only learn domain-invariant features but also enhance their generalization capabilities to new objects. Extensive experiments conducted on benchmark datasets demonstrate that our proposed approach surpasses state-of-the-art object detection methods by 6.8 mAP score in terms of detection accuracy on the Microsoft COCO dataset. Notably, the model size remains a compact 3.2M, enabling fast inference speeds, lower computational resource consumption, and enhanced resilience to domain variations. Adapt-Net represents a promising and efficient approach to object detection that combines accuracy with efficiency.
引用
收藏
页码:120788 / 120803
页数:16
相关论文
共 50 条
  • [31] Demo: Fast and Accurate Object Analysis at the Edge for Mobile Augmented Reality
    Liu, Qiang
    Huang, Siqi
    Han, Tao
    SEC 2017: 2017 THE SECOND ACM/IEEE SYMPOSIUM ON EDGE COMPUTING (SEC'17), 2017,
  • [32] AREDAPPS: Mobile Augmented Reality Development and Learning Framework Based on Augmented Reality Technology for Engineering Drawing Course
    Ali, Dayana Farzeeha
    Omar, Marlissa
    Sunar, Mohd Shahrizal
    Zaid, Norasykin Mohd
    Ibrahim, Nor Hasniza
    Surif, Johari
    INTELLIGENT TECHNOLOGIES FOR INTERACTIVE ENTERTAINMENT, INTETAIN 2021, 2022, 429 : 322 - 335
  • [33] A unified model sharing framework for moving object detection
    Chen, Yingying
    Wang, Jinqiao
    Xu, Min
    He, Xiangjian
    Lu, Hanqing
    SIGNAL PROCESSING, 2016, 124 : 72 - 80
  • [34] Unbalanced Optimal Transport: A Unified Framework for Object Detection
    De Plaen, Henri
    De Plaen, Pierre-Francois
    Suykens, Johan A. K.
    Proesmans, Marc
    Tuytelaars, Tinne
    Van Gool, Luc
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3198 - 3207
  • [35] Polyhedral object detection and pose estimation for augmented reality applications
    Shahrokni, A
    Vacchetti, L
    Lepetit, V
    Fua, P
    CA 2002: PROCEEDINGS OF THE COMPUTER ANIMATION 2002, 2002, : 65 - 69
  • [36] Integrating YOLO Object Detection with Augmented Reality for iOS Apps
    Mahurkar, Sagar
    2018 9TH IEEE ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), 2018, : 585 - 589
  • [37] A Unified Framework for Augmented Reality and Knowledge-Based Systems in Maintaining Aircraft
    Jo, Geun-Sik
    Oh, Kyeong-Jin
    Ha, Inay
    Lee, Kee-Sung
    Hong, Myung-Duk
    Neumann, Ulrich
    You, Suya
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 2990 - 2997
  • [38] LEAF plus AIO: Edge-Assisted Energy-Aware Object Detection for Mobile Augmented Reality
    Wang, Haoxin
    Kim, Baekgyu
    Xie, Jiang
    Han, Zhu
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (10) : 5933 - 5948
  • [39] Using mobile-based augmented reality and object detection for real-time Abalone growth monitoring
    Napier, Thomas
    Lee, Ickjai
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2023, 207
  • [40] Object detection for a mobile robot using mixed reality
    Chen, Hua
    Wulf, Oliver
    Wagner, Bernardo
    INTERACTIVE TECHNOLOGIES AND SOCIOTECHNICAL SYSTEMS, 2006, 4270 : 466 - 475