Self-supervised Representation Learning for Fine Grained Human Hand Action Recognition in Industrial Assembly Lines

被引:0
|
作者
Sturm, Fabian [1 ,2 ]
Sathiyababu, Rahul [1 ]
Allipilli, Harshitha [1 ]
Hergenroether, Elke [2 ]
Siegel, Melanie [2 ]
机构
[1] Bosch Rexroth AG, Lise Meitner Str 4, D-89081 Ulm, Germany
[2] Univ Appl Sci Darmstadt, Schoefferstr 3, D-64295 Darmstadt, Germany
关键词
Self-Supervised Learning; Human Action Recognition; Industrial Vision;
D O I
10.1007/978-3-031-47969-4_14
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Humans are still indispensable on industrial assembly lines, but in the event of an error, they need support from intelligent systems. In addition to the objects to be observed, it is equally important to understand the fine-grained hand movements of a human to be able to track the entire process. However, these deep learning based hand action recognition methods are very label intensive, which cannot be offered by all industrial companies due to the associated costs. This work therefore presents a self-supervised learning approach for industrial assembly processes that allows a spatio-temporal transformer architecture to be pre-trained on a variety of information from real-world video footage of daily life. Subsequently, this deep learning model is adapted to the industrial assembly task at hand using only a few labels. It is shown which known real-world datasets are best suited for representation learning of these hand actions in a regression task, and to what extent they optimize the subsequent supervised trained classification task.
引用
收藏
页码:172 / 184
页数:13
相关论文
共 50 条
  • [1] Semi-Supervised Learning Approach for Fine Grained Human Hand Action Recognition in Industrial Assembly
    Sturm, Fabian
    Sathiyababu, Rahul
    Hergenroether, Elke
    Siegel, Melanie
    [J]. Computer Science Research Notes, 2023, 31 (1-2): : 340 - 350
  • [2] Learning Common Rationale to Improve Self-Supervised Representation for Fine-Grained Visual Recognition Problems
    Shu, Yangyang
    van den Hengel, Anton
    Liu, Lingqiao
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11392 - 11401
  • [3] ATOM: Self-supervised human action recognition using atomic motion representation learning
    Degardin, Bruno
    Lopes, Vasco
    Proenca, Hugo
    [J]. IMAGE AND VISION COMPUTING, 2023, 137
  • [4] Self-supervised learning for visual tracking and recognition of human hand
    Wu, Y
    Huang, TS
    [J]. SEVENTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-2001) / TWELFTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-2000), 2000, : 243 - 248
  • [5] Spatiotemporal consistency enhancement self-supervised representation learning for action recognition
    Bi, Shuai
    Hu, Zhengping
    Zhao, Mengyao
    Li, Shufang
    Sun, Zhe
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (04) : 1485 - 1492
  • [6] Spatiotemporal consistency enhancement self-supervised representation learning for action recognition
    Shuai Bi
    Zhengping Hu
    Mengyao Zhao
    Shufang Li
    Zhe Sun
    [J]. Signal, Image and Video Processing, 2023, 17 : 1485 - 1492
  • [7] Contrastive self-supervised representation learning without negative samples for multimodal human action recognition
    Yang, Huaigang
    Ren, Ziliang
    Yuan, Huaqiang
    Xu, Zhenyu
    Zhou, Jun
    [J]. FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [8] SELF SUPERVISED DEEP REPRESENTATION LEARNING FOR FINE-GRAINED BODY PART RECOGNITION
    Zhang, Pengyue
    Wang, Fusheng
    Zheng, Yefeng
    [J]. 2017 IEEE 14TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2017), 2017, : 578 - 582
  • [9] Self-supervised representation learning for surgical activity recognition
    Paysan, Daniel
    Haug, Luis
    Bajka, Michael
    Oelhafen, Markus
    Buhmann, Joachim M.
    [J]. INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2021, 16 (11) : 2037 - 2044
  • [10] Self-Supervised ECG Representation Learning for Emotion Recognition
    Sarkar, Pritam
    Etemad, Ali
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (03) : 1541 - 1554