Audio-Driven Laughter Behavior Controller

被引:6
|
作者
Ding, Yu [1 ]
Huang, Jing [2 ]
Pelachaud, Catherine [3 ]
机构
[1] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
[2] Zhejiang Gongshang Univ, Sch Informat & Elect Engn, Hangzhou 310018, Zhejiang, Peoples R China
[3] Univ Paris 06, CNRS, ISIR, F-75005 Paris, France
关键词
Laughter; audio-driven; data-driven; animation synthesis; continuous-state; Kalman filter; prosody; nonverbal behaviors; virtual character; statistical framework;
D O I
10.1109/TAFFC.2017.2754365
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It has been well documented that laughter is an important communicative and expressive signal in face-to-face conversations. Our work aims at building a laughter behavior controller for a virtual character which is able to generate upper body animations from laughter audio given as input. This controller relies on the tight correlations between laughter audio and body behaviors. A unified continuous-state statistical framework, inspired by Kalman filter, is proposed to learn the correlations between laughter audio and head/torso behavior from a recorded laughter human dataset. Due to the lack of shoulder behavior data in the recorded human dataset, a rule-based method is defined to model the correlation between laughter audio and shoulder behavior. In the synthesis step, these characterized correlations are rendered in the animation of a virtual character. To validate our controller, a subjective evaluation is conducted where participants viewed the videos of a laughing virtual character. It compares the animations of a virtual character using our controller and a state of the art method. The evaluation results show that the laughter animations computed with our controller are perceived as more natural, expressing amusement more freely and appearing more authentic than with the state of the art method.
引用
收藏
页码:546 / 558
页数:13
相关论文
共 50 条
  • [41] SoundToons: Exemplar-Based Authoring of Interactive Audio-Driven Animation Sprites
    Chong, Toby
    Shin, Hijung Valentina
    Aneja, Deepali
    Igarashi, Takeo
    PROCEEDINGS OF 2023 28TH ANNUAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2023, 2023, : 710 - 722
  • [42] Taming Diffusion Models for Audio-Driven Co-Speech Gesture Generation
    Zhu, Lingting
    Liu, Xian
    Liu, Xuanyu
    Qian, Rui
    Liu, Ziwei
    Yu, Lequan
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10544 - 10553
  • [43] Semantic-Aware Implicit Neural Audio-Driven Video Portrait Generation
    Liu, Xian
    Xu, Yinghao
    Wu, Qianyi
    Zhou, Hang
    Wu, Wayne
    Zhou, Bolei
    COMPUTER VISION, ECCV 2022, PT XXXVII, 2022, 13697 : 106 - 125
  • [44] Audio-driven Talking Head Generation with Transformer and 3D Morphable Model
    Huang, Ricong
    Zhong, Weizhi
    Li, Guanbin
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 7035 - 7039
  • [45] Audio2Moves: Two-Level Hierarchical Framework for Audio-Driven Human Motion Synthesis
    Yanbo Cheng
    Nada Elmasry
    Yingying Wang
    SN Computer Science, 6 (5)
  • [46] Voice2Face: Audio-driven Facial and Tongue Rig Animations with cVAEs
    Aylagas, Monica Villanueva
    Leon, Hector Anadon
    Teye, Mattias
    Tollmar, Konrad
    COMPUTER GRAPHICS FORUM, 2022, 41 (08) : 255 - 265
  • [47] MergeTalk: Audio-Driven Talking Head Generation From Single Image With Feature Merge
    Gao, Jian
    Shu, Chang
    Zheng, Ximin
    Lu, Zheng
    Bao, Nengsheng
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1850 - 1854
  • [48] DiffuseStyleGesture: Stylized Audio-Driven Co-Speech Gesture Generation with Diffusion Models
    Yang, Sicheng
    Wu, Zhiyong
    Li, Minglei
    Zhang, Zhensong
    Hao, Lei
    Bao, Weihong
    Cheng, Ming
    Xiao, Long
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 5860 - 5868
  • [49] Audio-Driven Facial Animation by Joint End-to-End Learning of Pose and Emotion
    Karras, Tero
    Aila, Timo
    Laine, Samuli
    Herva, Antti
    Lehtinen, Jaakko
    ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (04):
  • [50] AVForensics: Audio-driven Deepfake Video Detection with Masking Strategy in Self-supervision
    Zhu Yizhe
    Gao Jialin
    Zhou Xi
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 162 - 171