Enhancing Recognition of Human-Object Interaction from Visual Data Using Egocentric Wearable Camera

被引:0
|
作者
Hamid, Danish [1 ]
Ul Haq, Muhammad Ehatisham [1 ]
Yasin, Amanullah [1 ]
Murtaza, Fiza [1 ]
Azam, Muhammad Awais [2 ]
机构
[1] Air Univ, Fac Comp & Artificial Intelligence FCAI, Dept Creat Technol, Islamabad 44000, Pakistan
[2] Whitecliffe, Technol & Innovat Res Grp, Sch Informat Technol, Wellington 6145, New Zealand
关键词
egocentric; hand pose; human-object interaction; machine learning; object recognition; wearable camera;
D O I
10.3390/fi16080269
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection and human action recognition have great significance in many real-world applications. Understanding how a human being interacts with different objects, i.e., human-object interaction, is also crucial in this regard since it enables diverse applications related to security, surveillance, and immersive reality. Thus, this study explored the potential of using a wearable camera for object detection and human-object interaction recognition, which is a key technology for the future Internet and ubiquitous computing. We propose a system that uses an egocentric camera view to recognize objects and human-object interactions by analyzing the wearer's hand pose. Our novel idea leverages the hand joint data of the user, which were extracted from the egocentric camera view, for recognizing different objects and related interactions. Traditional methods for human-object interaction rely on a third-person, i.e., exocentric, camera view by extracting morphological and color/texture-related features, and thus, often fall short when faced with occlusion, camera variations, and background clutter. Moreover, deep learning-based approaches in this regard necessitate substantial data for training, leading to a significant computational overhead. Our proposed approach capitalizes on hand joint data captured from an egocentric perspective, offering a robust solution to the limitations of traditional methods. We propose a machine learning-based innovative technique for feature extraction and description from 3D hand joint data by presenting two distinct approaches: object-dependent and object-independent interaction recognition. The proposed method offered advantages in computational efficiency compared with deep learning methods and was validated using the publicly available HOI4D dataset, where it achieved a best-case average F1-score of 74%. The proposed system paves the way for intuitive human-computer collaboration within the future Internet, enabling applications like seamless object manipulation and natural user interfaces for smart devices, human-robot interactions, virtual reality, and augmented reality.
引用
下载
收藏
页数:17
相关论文
共 50 条
  • [21] Spatio-Temporal Interaction Graph Parsing Networks for Human-Object Interaction Recognition
    Wang, Ning
    Zhu, Guangming
    Zhang, Liang
    Shen, Peiyi
    Li, Hongsheng
    Hua, Cong
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4985 - 4993
  • [22] Dangerous Human Event Understanding using Human-Object Interaction Model
    Xu, Zhaozhuo
    Tian, Yuan
    Hu, Xinjue
    Pu, Fangling
    2015 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, COMMUNICATIONS AND COMPUTING (ICSPCC), 2015, : 150 - 154
  • [23] STIT: Spatio-Temporal Interaction Transformers for Human-Object Interaction Recognition in Videos
    Almushyti, Muna
    Li, Frederick W. B.
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 3287 - 3294
  • [24] Exploring the synergy between textual identity and visual signals in human-object interaction
    An, Pinzhu
    Tan, Zhi
    IMAGE AND VISION COMPUTING, 2024, 151
  • [25] Medication and Meal Intake Monitoring using Human-Object Interaction
    Seint, Pann Thinzar
    Zin, Thi Thi
    Yokota, Mitsuhiro
    2018 IEEE 7TH GLOBAL CONFERENCE ON CONSUMER ELECTRONICS (GCCE 2018), 2018, : 399 - 400
  • [26] Visual-Semantic Graph Attention Networks for Human-Object Interaction Detection
    Liang, Zhijun
    Liu, Junfa
    Guan, Yisheng
    Rojas, Juan
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE-ROBIO 2021), 2021, : 1441 - 1447
  • [27] HOIMotion: Forecasting Human Motion during Human-Object Interactions Using Egocentric 3D Object Bounding Boxes
    Hu, Zhiming
    Yin, Zheming
    Haeufle, Daniel
    Schmitt, Syn
    Bulling, Andreas
    IEEE Transactions on Visualization and Computer Graphics, 2024, 30 (11) : 7375 - 7385
  • [28] Discriminative Orderlet Mining for Real-Time Recognition of Human-Object Interaction
    Yu, Gang
    Liu, Zicheng
    Yuan, Junsong
    COMPUTER VISION - ACCV 2014, PT V, 2015, 9007 : 50 - 65
  • [29] Scaling Human-Object Interaction Recognition through Zero-Shot Learning
    Shen, Liyue
    Yeung, Serena
    Hoffman, Judy
    Mori, Greg
    Li Fei-Fei
    2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, : 1568 - 1576
  • [30] Knowledge-Based Role Recognition by Using Human-Object Interaction and Spatio-Temporal Analysis
    Yang, Chule
    Zeng, Yijie
    Yue, Yufeng
    Siritanawan, Prarinya
    Zhang, Jun
    Wang, Danwei
    2017 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE ROBIO 2017), 2017, : 159 - 164