Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions

被引:250
|
作者
Bambach, Sven [1 ]
Lee, Stefan [1 ]
Crandall, David J. [1 ]
Yu, Chen [2 ]
机构
[1] Indiana Univ, Sch Informat & Comp, Bloomington, IN 47405 USA
[2] Indiana Univ, Psychol & Brain Sci, Bloomington, IN 47405 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/ICCV.2015.226
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hands appear very often in egocentric video, and their appearance and pose give important cues about what people are doing and what they are paying attention to. But existing work in hand detection has made strong assumptions that work well in only simple scenarios, such as with limited interaction with other people or in lab settings. We develop methods to locate and distinguish between hands in egocentric video using strong appearance models with Convolutional Neural Networks, and introduce a simple candidate region generation approach that outperforms existing techniques at a fraction of the computational cost. We show how these high-quality bounding boxes can be used to create accurate pixelwise hand regions, and as an application, we investigate the extent to which hand segmentation alone can distinguish between different activities. We evaluate these techniques on a new dataset of 48 first-person videos of people interacting in realistic environments, with pixel-level ground truth for over 15,000 hand instances.
引用
收藏
页码:1949 / 1957
页数:9
相关论文
共 50 条
  • [41] Mining intricate temporal rules for recognizing complex activities of daily living under uncertainty
    Liu, Li
    Wang, Shu
    Peng, Yuxin
    Huang, Zigang
    Liu, Ming
    Hu, Bin
    [J]. PATTERN RECOGNITION, 2016, 60 : 1015 - 1028
  • [42] Recognizing complex instrumental activities of daily living using scene information and fuzzy logic
    Banerjee, Tanvi
    Keller, James M.
    Popescu, Mihail
    Skubic, Marjorie
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2015, 140 : 68 - 82
  • [43] Multifunctional Self-Powered Sensors Integrated on a Robot Hand for Detecting Temperature-Pressure Stimuli and Recognizing Objects
    Qi, Xiangyu
    Wang, Linglu
    Li, Chuanbo
    Wang, Yang
    [J]. ACS Applied Materials and Interfaces, 2024, 16 (40): : 54475 - 54484
  • [44] A Graph-Based Approach to Recognizing Complex Human Object Interactions in Sequential Data
    Ghadi, Yazeed Yasin
    Waheed, Manahil
    Gochoo, Munkhjargal
    Alsuhibany, Suliman A.
    Chelloug, Samia Allaoua
    Jalal, Ahmad
    Park, Jeongmin
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (10):
  • [45] Using Complex-Valued Levenberg-Marquardt Algorithm for Learning and Recognizing Various Hand Gestures
    Hafiz, Abdul Rahman
    Amin, Md Faijul
    Murase, Kazuyuki
    [J]. 2012 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2012,
  • [46] PARCIV: Recognizing physical activities having complex interclass variations using semantic data of smartphone
    Usman Sarwar, Muhammad
    Rehman Javed, Abdul
    Kulsoom, Farzana
    Khan, Suleman
    Tariq, Usman
    Kashif Bashir, Ali
    [J]. SOFTWARE-PRACTICE & EXPERIENCE, 2021, 51 (03): : 532 - 549
  • [47] VISUALIZING THE INVISIBLE HAND OF MARKETS: SIMULATING COMPLEX DYNAMIC ECONOMIC INTERACTIONS
    Jaffe, Klaus
    [J]. INTELLIGENT SYSTEMS IN ACCOUNTING FINANCE & MANAGEMENT, 2015, 22 (02): : 115 - 132
  • [48] Deep-HAR: an ensemble deep learning model for recognizing the simple, complex, and heterogeneous human activities
    Prabhat Kumar
    S Suresh
    [J]. Multimedia Tools and Applications, 2023, 82 : 30435 - 30462
  • [49] Parent–child pair design for detecting gene–environment interactions in complex diseases
    Yuan-De Tan
    Myriam Fornage
    Varghese George
    Hongyan Xu
    [J]. Human Genetics, 2007, 121 : 745 - 757
  • [50] Deep-HAR: an ensemble deep learning model for recognizing the simple, complex, and heterogeneous human activities
    Kumar, Prabhat
    Suresh, S.
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (20) : 30435 - 30462