Diffusion-Based Unsupervised Pre-training for Automated Recognition of Vitality Forms

被引:0
|
作者
Canovi, Noemi [1 ]
Montagna, Federico [1 ]
Niewiadomski, Radoslaw [2 ]
Sciutti, Alessandra [3 ]
Di Cesare, Giuseppe [3 ,4 ]
Beyan, Cigdem [5 ]
机构
[1] Univ Trento, Dep Informat Engn & Comp Sci, Trento, Italy
[2] Univ Genoa, Dept Informat Bioengn Robot & Syst Engn, Genoa, Italy
[3] Ist Italiano Tecnol, CONTACT Unit, Genoa, Italy
[4] Univ Parma, Dept Med & Surg, Parma, Italy
[5] Univ Verona, Dept Comp Sci, Verona, Italy
基金
欧洲研究理事会;
关键词
Vitality forms; nonverbal communication; unsupervised pre-training; diffusion models; autoencoders; gestures; actions; trajectory; EXPRESSION;
D O I
10.1145/3656650.3656689
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Social communication involves interpreting nonverbal behaviors, detecting and anticipating others' actions and intentions. Actions convey not only the goal and motor intention but also the form, i.e., variations in action execution. These variations, termed vitality forms, communicate attitudes during interactions, such as being gentle, calm, vigorous, and rude. Automatic vitality form recognition may have several applications in social robotics, social skills training, and therapy, yet it remains a rarely studied topic. This paper introduces an unsupervised pre-training approach that utilizes 2D-body key point trajectories as input and employs diffusion models to derive more effective features for representing these trajectories. The features learned from the diffusion model's encoder are utilized to train a multilayer perceptron for vitality form recognition. Experimental analysis showcases the superior performance of the proposed method not only across various videos but also for action classes not encountered during training.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Unsupervised Pre-training Classifier Based on Restricted Boltzmann Machine with Imbalanced Data
    Fu, Xiaoyang
    [J]. SMART COMPUTING AND COMMUNICATION, SMARTCOM 2016, 2017, 10135 : 102 - 110
  • [22] Lottery Hypothesis based Unsupervised Pre-training for Model Compression in Federated Learning
    Itahara, Sohei
    Nishio, Takayuki
    Morikura, Masahiro
    Yamamoto, Koji
    [J]. 2020 IEEE 92ND VEHICULAR TECHNOLOGY CONFERENCE (VTC2020-FALL), 2020,
  • [23] Diffusion-based network for unsupervised landmark detection
    Wu, Tao
    Wang, Kai
    Tang, Chuanming
    Zhang, Jianlin
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 292
  • [24] Unleashing the Transferability Power of Unsupervised Pre-Training for Emotion Recognition in Masked and Unmasked Facial Images
    D'Inca, Moreno
    Beyan, Cigdem
    Niewiadomski, Radoslaw
    Barattin, Simone
    Sebe, Nicu
    [J]. IEEE ACCESS, 2023, 11 : 90876 - 90890
  • [25] Improving Transformer-based Speech Recognition with Unsupervised Pre-training and Multi-task Semantic Knowledge Learning
    Li, Song
    Li, Lin
    Hong, Qingyang
    Liu, Lingling
    [J]. INTERSPEECH 2020, 2020, : 5006 - 5010
  • [26] In Defense of Image Pre-Training for Spatiotemporal Recognition
    Li, Xianhang
    Wang, Huiyu
    Wei, Chen
    Mei, Jieru
    Yuille, Alan
    Zhou, Yuyin
    Xie, Cihang
    [J]. COMPUTER VISION, ECCV 2022, PT XXV, 2022, 13685 : 675 - 691
  • [27] An Empirical Study on Unsupervised Pre-training Approaches in Regression Problems
    Saikia, Pallabi
    Baruah, Rashmi Dutta
    [J]. 2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 342 - 349
  • [28] Depression recognition using voice-based pre-training model
    Huang, Xiangsheng
    Wang, Fang
    Gao, Yuan
    Liao, Yilong
    Zhang, Wenjing
    Zhang, Li
    Xu, Zhenrong
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01):
  • [29] Unsupervised pre-training of graph transformers on patient population graphs
    Pellegrini, Chantal
    Navab, Nassir
    Kazi, Anees
    [J]. MEDICAL IMAGE ANALYSIS, 2023, 89
  • [30] A Multilingual Framework Based on Pre-training Model for Speech Emotion Recognition
    Zhang, Zhaohang
    Zhang, Xiaohui
    Guo, Min
    Zhang, Wei-Qiang
    Li, Ke
    Huang, Yukai
    [J]. 2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 750 - 755