Robotic Facial Emotion Transfer Network Based on Transformer Framework and B-spline Smoothing Constraint

被引:0
|
作者
Huang Z. [1 ,2 ]
Ren F. [3 ]
Hu M. [2 ]
Liu J. [1 ]
机构
[1] School of Electronic Engineering and Intelligent Manufacturing, Anqing Normal University, Anqing
[2] School of Computer Science and Information, Hefei University of Technology, Hefei
[3] Faculty of Engineering, University of Tokushima, Tokushima
来源
Jiqiren/Robot | 2023年 / 45卷 / 04期
关键词
cubic B-spline smoothing constraint; facial emotion transfer; humanoid robot; inter-domain cooperative attention; intra-domain deformation attention;
D O I
10.13973/j.cnki.robot.220351
中图分类号
学科分类号
摘要
To improve the spatial-temporal consistency of facial emotion transfer and reduce the influence of mechanical motion constraints for humanoid robot, a robotic facial emotion transformer (RFEFormer) network based on Transformer framework and B-spline smoothing constraint is proposed. The RFEFormer network consists of facial deformation encode subnet and actuation sequence generation subnet. In facial deformation encode subnet, an intra-frame spatial attention module, which is constructed based on dual mechanisms of intra-domain deformation attention and inter-domain cooperative attention, is embedded into Transformer encoder to represent the intra-frame spatial information of different levels and granularities. In actuation sequence generation subnet, a Transformer decoder, which accomplishes cross attention of facial spatio-temporal sequence and history motor actuation sequence, is addressed for multi-step prediction of future motor drive sequence. Moreover, a cubic B-spline smoothing constraint is introduced to realize the warping of prediction sequence. The experimental results show that the motor actuation deviation, the facial deformation fidelity and motor motion smoothness of the RFEFormer network is 3.21%, 89.48% and 90.63%, respectively. Furthermore, the frame rate of the real-time facial emotion transfer is greater than 25 frames per second. Compared with the related methods, the proposed RFEFormer network not only satisfies the real-time performance, but also improves the time sequence-based indexes such as fidelity and smoothness, which are more sensitive and concerned by human senses. © 2023 Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:395 / 408
页数:13
相关论文
共 26 条
  • [1] Abdollahi H, Mahoor M, Zandie R, Et al., Artificial emotional intelligence in socially assistive robots for older adults: A pilot study[J], IEEE Transactions on Affective Computing, (2022)
  • [2] Stock-Homburg R., Survey of emotions in human-robot interactions: Perspectives from robotic psychology on 20 years of research, International Journal of Social Robotics, 14, 3, pp. 389-411, (2022)
  • [3] Chiang A H, Trimi S, Lo Y J., Emotion and service quality of anthropomorphic robots[J], Technological Forecasting and Social Change, 177, 4, (2022)
  • [4] Geiger A R, Balas B., Robot face memorability is affected by uncanny appearance[J], Computers in Human Behavior Reports, 4, 11, (2021)
  • [5] Lalitharatne T D, Tan J, He L, Et al., MorphFace: A hybrid morphable face for a Robopatient, IEEE Robotics and Automation Letters, 6, 2, pp. 643-650, (2021)
  • [6] Sonmez E B, Han H, Karadeniz O, Et al., EMRES: A new emotional respondent robot, IEEE Transactions on Cognitive and Developmental Systems, 14, 2, pp. 772-780, (2022)
  • [7] Machaen Z, Martin L, Rosales J H., Bio-inspired cognitive model of motor learning by imitation, Cognitive Systems Research, 66, 3, pp. 134-149, (2021)
  • [8] Huang Z, Ren F J, Hu M, Et al., A real-time expression mimicking method for humanoid robot based on dual LSTM fusion, Robot, 41, 2, pp. 137-146, (2019)
  • [9] Hyung H J, Yoon H U, Choi D W, Et al., Optimizing android facial expressions using genetic algorithms, Applied Sciences, 9, 16, pp. 3379-3385, (2019)
  • [10] Ren F J, Huang Z., Automatic facial expression learning method based on humanoid robot XIN-REN, IEEE Transactions on Human-Machine Systems, 46, 6, pp. 810-821, (2016)