Enhancing Recognition of Human-Object Interaction from Visual Data Using Egocentric Wearable Camera

被引:0
|
作者
Hamid, Danish [1 ]
Ul Haq, Muhammad Ehatisham [1 ]
Yasin, Amanullah [1 ]
Murtaza, Fiza [1 ]
Azam, Muhammad Awais [2 ]
机构
[1] Air Univ, Fac Comp & Artificial Intelligence FCAI, Dept Creat Technol, Islamabad 44000, Pakistan
[2] Whitecliffe, Technol & Innovat Res Grp, Sch Informat Technol, Wellington 6145, New Zealand
关键词
egocentric; hand pose; human-object interaction; machine learning; object recognition; wearable camera;
D O I
10.3390/fi16080269
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection and human action recognition have great significance in many real-world applications. Understanding how a human being interacts with different objects, i.e., human-object interaction, is also crucial in this regard since it enables diverse applications related to security, surveillance, and immersive reality. Thus, this study explored the potential of using a wearable camera for object detection and human-object interaction recognition, which is a key technology for the future Internet and ubiquitous computing. We propose a system that uses an egocentric camera view to recognize objects and human-object interactions by analyzing the wearer's hand pose. Our novel idea leverages the hand joint data of the user, which were extracted from the egocentric camera view, for recognizing different objects and related interactions. Traditional methods for human-object interaction rely on a third-person, i.e., exocentric, camera view by extracting morphological and color/texture-related features, and thus, often fall short when faced with occlusion, camera variations, and background clutter. Moreover, deep learning-based approaches in this regard necessitate substantial data for training, leading to a significant computational overhead. Our proposed approach capitalizes on hand joint data captured from an egocentric perspective, offering a robust solution to the limitations of traditional methods. We propose a machine learning-based innovative technique for feature extraction and description from 3D hand joint data by presenting two distinct approaches: object-dependent and object-independent interaction recognition. The proposed method offered advantages in computational efficiency compared with deep learning methods and was validated using the publicly available HOI4D dataset, where it achieved a best-case average F1-score of 74%. The proposed system paves the way for intuitive human-computer collaboration within the future Internet, enabling applications like seamless object manipulation and natural user interfaces for smart devices, human-robot interactions, virtual reality, and augmented reality.
引用
下载
收藏
页数:17
相关论文
共 50 条
  • [41] Geometric Features Informed Multi-person Human-Object Interaction Recognition in Videos
    Qiao, Tanqiu
    Men, Qianhui
    Li, Frederick W. B.
    Kubotani, Yoshiki
    Morishima, Shigeo
    Hubert, P. H. Shum
    COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 474 - 491
  • [42] Detection of Anomalous Behavior of Manufacturing Workers Using Deep Learning-Based Recognition of Human-Object Interaction
    Rijayanti, Rita
    Hwang, Mintae
    Jin, Kyohong
    APPLIED SCIENCES-BASEL, 2023, 13 (15):
  • [43] Human-object interaction recognition based on interactivity detection and multi-feature fusion
    Xia, Limin
    Ding, Xiaoyue
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2024, 27 (02): : 1169 - 1183
  • [44] Human-object interaction detection via interactive visual-semantic graph learning
    Tongtong Wu
    Fuqing Duan
    Liang Chang
    Ke Lu
    Science China Information Sciences, 2022, 65
  • [45] HOI4D: A 4D Egocentric Dataset for Category-Level Human-Object Interaction
    Liu, Yunze
    Liu, Yun
    Jiang, Che
    Lyu, Kangbo
    Wan, Weikang
    Shen, Hao
    Liang, Boqiang
    Fu, Zhoujie
    Wang, He
    Yi, Li
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 20981 - 20990
  • [46] The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain
    Ragusa, Francesco
    Furnari, Antonino
    Livatino, Salvatore
    Farinella, Giovanni Maria
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1568 - 1577
  • [47] From detection to understanding: A survey on representation learning for human-object interaction
    Luo, Tianlun
    Guan, Steven
    Yang, Rui
    Smith, Jeremy
    NEUROCOMPUTING, 2023, 543
  • [48] Attribute Based Affordance Detection from Human-Object Interaction Images
    Hassan, Mahmudul
    Dharmaratne, Anuja
    IMAGE AND VIDEO TECHNOLOGY - PSIVT 2015 WORKSHOPS, 2016, 9555 : 220 - 232
  • [49] Visual SLAM With BoPLW Pairs Using Egocentric Stereo Camera for Wearable-Assisted Substation Inspection
    Qian, Kun
    Zhao, Wei
    Li, Kai
    Ma, Xudong
    Yu, Hai
    IEEE SENSORS JOURNAL, 2020, 20 (03) : 1630 - 1641
  • [50] Designing smart living objects - Enhancing vs. distracting traditional human-object interaction
    Chi, Pei-yu
    Chen, Jen-hao
    Liu, Shih-yen
    Chu, Hao-hua
    HUMAN-COMPUTER INTERACTION, PT 2, PROCEEDINGS, 2007, 4551 : 788 - +