Enhancing Recognition of Human-Object Interaction from Visual Data Using Egocentric Wearable Camera

被引:0
|
作者
Hamid, Danish [1 ]
Ul Haq, Muhammad Ehatisham [1 ]
Yasin, Amanullah [1 ]
Murtaza, Fiza [1 ]
Azam, Muhammad Awais [2 ]
机构
[1] Air Univ, Fac Comp & Artificial Intelligence FCAI, Dept Creat Technol, Islamabad 44000, Pakistan
[2] Whitecliffe, Technol & Innovat Res Grp, Sch Informat Technol, Wellington 6145, New Zealand
关键词
egocentric; hand pose; human-object interaction; machine learning; object recognition; wearable camera;
D O I
10.3390/fi16080269
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection and human action recognition have great significance in many real-world applications. Understanding how a human being interacts with different objects, i.e., human-object interaction, is also crucial in this regard since it enables diverse applications related to security, surveillance, and immersive reality. Thus, this study explored the potential of using a wearable camera for object detection and human-object interaction recognition, which is a key technology for the future Internet and ubiquitous computing. We propose a system that uses an egocentric camera view to recognize objects and human-object interactions by analyzing the wearer's hand pose. Our novel idea leverages the hand joint data of the user, which were extracted from the egocentric camera view, for recognizing different objects and related interactions. Traditional methods for human-object interaction rely on a third-person, i.e., exocentric, camera view by extracting morphological and color/texture-related features, and thus, often fall short when faced with occlusion, camera variations, and background clutter. Moreover, deep learning-based approaches in this regard necessitate substantial data for training, leading to a significant computational overhead. Our proposed approach capitalizes on hand joint data captured from an egocentric perspective, offering a robust solution to the limitations of traditional methods. We propose a machine learning-based innovative technique for feature extraction and description from 3D hand joint data by presenting two distinct approaches: object-dependent and object-independent interaction recognition. The proposed method offered advantages in computational efficiency compared with deep learning methods and was validated using the publicly available HOI4D dataset, where it achieved a best-case average F1-score of 74%. The proposed system paves the way for intuitive human-computer collaboration within the future Internet, enabling applications like seamless object manipulation and natural user interfaces for smart devices, human-robot interactions, virtual reality, and augmented reality.
引用
下载
收藏
页数:17
相关论文
共 50 条
  • [31] Combining monoSLAM with object recognition for scene augmentation using a wearable camera
    Castle, R. O.
    Klein, G.
    Murray, D. W.
    IMAGE AND VISION COMPUTING, 2010, 28 (11) : 1548 - 1556
  • [32] SCRATCHHOI: TRAINING HUMAN-OBJECT INTERACTION DETECTORS FROM SCRATCH
    Lim, Junyi
    Baskaran, Vishnu Monn
    Lim, Joanne MunYee
    Sutopo, Ricky
    Wong, KokSheik
    Tistarelli, Massimo
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1690 - 1694
  • [33] Interaction Replica: Tracking Human-Object Interaction and Scene Changes From Human Motion
    Guzov, Vladimir
    Chibane, Julian
    Marin, Riccardo
    He, Yannan
    Saracoglu, Yunus
    Sattler, Torsten
    Pons-Moll, Gerard
    2024 INTERNATIONAL CONFERENCE IN 3D VISION, 3DV 2024, 2024, : 1006 - 1016
  • [34] Observing Human-Object Interactions: Using Spatial and Functional Compatibility for Recognition
    Gupta, Abhinav
    Kembhavi, Aniruddha
    Davis, Larry S.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (10) : 1775 - 1789
  • [35] A SIMILARITY MEASURE FOR ANALYZING HUMAN ACTIVITIES USING HUMAN-OBJECT INTERACTION CONTEXT
    Amiri, S. Mohsen
    Pourazad, Mahsa T.
    Nasiopoulos, Panos
    Leung, Victor C. M.
    2014 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2014, : 2368 - 2372
  • [36] Human Daily Activity Recognition for Healthcare Using Wearable and Visual Sensing Data
    Liu, Xi
    Liu, Lei
    Simske, Steven J.
    Liu, Jerry
    2016 IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI), 2016, : 24 - 31
  • [37] Human-object interaction detection via interactive visual-semantic graph learning
    Tongtong WU
    Fuqing DUAN
    Liang CHANG
    Ke LU
    Science China(Information Sciences), 2022, 65 (06) : 81 - 82
  • [38] Human-object interaction detection via interactive visual-semantic graph learning
    Wu, Tongtong
    Duan, Fuqing
    Chang, Liang
    Lu, Ke
    SCIENCE CHINA-INFORMATION SCIENCES, 2022, 65 (06)
  • [39] Language-guided graph parsing attention network for human-object interaction recognition
    Li, Qiyue
    Xie, Xuemei
    Zhang, Jin
    Shi, Guangming
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 89
  • [40] Scaling Human-Object Interaction Recognition in the Video through Zero-Shot Learning
    Maraghi, Vali Ollah
    Faez, Karim
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2021, 2021