Audio-Driven Laughter Behavior Controller

被引:6
|
作者
Ding, Yu [1 ]
Huang, Jing [2 ]
Pelachaud, Catherine [3 ]
机构
[1] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
[2] Zhejiang Gongshang Univ, Sch Informat & Elect Engn, Hangzhou 310018, Zhejiang, Peoples R China
[3] Univ Paris 06, CNRS, ISIR, F-75005 Paris, France
关键词
Laughter; audio-driven; data-driven; animation synthesis; continuous-state; Kalman filter; prosody; nonverbal behaviors; virtual character; statistical framework;
D O I
10.1109/TAFFC.2017.2754365
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It has been well documented that laughter is an important communicative and expressive signal in face-to-face conversations. Our work aims at building a laughter behavior controller for a virtual character which is able to generate upper body animations from laughter audio given as input. This controller relies on the tight correlations between laughter audio and body behaviors. A unified continuous-state statistical framework, inspired by Kalman filter, is proposed to learn the correlations between laughter audio and head/torso behavior from a recorded laughter human dataset. Due to the lack of shoulder behavior data in the recorded human dataset, a rule-based method is defined to model the correlation between laughter audio and shoulder behavior. In the synthesis step, these characterized correlations are rendered in the animation of a virtual character. To validate our controller, a subjective evaluation is conducted where participants viewed the videos of a laughing virtual character. It compares the animations of a virtual character using our controller and a state of the art method. The evaluation results show that the laughter animations computed with our controller are perceived as more natural, expressing amusement more freely and appearing more authentic than with the state of the art method.
引用
收藏
页码:546 / 558
页数:13
相关论文
共 50 条
  • [1] An audio-driven dancing avatar
    Ofli, Ferda
    Demir, Yasemin
    Yemez, Yucel
    Erzin, Engin
    Tekalp, A. Murat
    Balci, Koray
    Kizoglu, Idil
    Akarun, Lale
    Canton-Ferrer, Cristian
    Tilmanne, Joelle
    Bozkurt, Elif
    Erdem, A. Tanju
    JOURNAL ON MULTIMODAL USER INTERFACES, 2008, 2 (02) : 93 - 103
  • [2] An audio-driven dancing avatar
    Ferda Ofli
    Yasemin Demir
    Yücel Yemez
    Engin Erzin
    A. Murat Tekalp
    Koray Balcı
    İdil Kızoğlu
    Lale Akarun
    Cristian Canton-Ferrer
    Joëlle Tilmanne
    Elif Bozkurt
    A. Tanju Erdem
    Journal on Multimodal User Interfaces, 2008, 2 : 93 - 103
  • [3] Photorealistic Audio-driven Video Portraits
    Wen, Xin
    Wang, Miao
    Richardt, Christian
    Chen, Ze-Yin
    Hu, Shi-Min
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (12) : 3457 - 3466
  • [4] Audio-Driven Emotional Video Portraits
    Ji, Xinya
    Zhou, Hang
    Wang, Kaisiyuan
    Wu, Wayne
    Loy, Chen Change
    Cao, Xun
    Xu, Feng
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14075 - 14084
  • [5] Audio-Driven Multimedia Content Authentication as a Service
    Vryzas, Nikolaos
    Katsaounidou, Anastasia
    Kotsakis, Rigas
    Dimoulas, Charalampos
    Kalliris, George
    146TH AES CONVENTION, 2019,
  • [6] Audio-Driven Talking Face Generation: A Review
    Liu, Shiguang
    JOURNAL OF THE AUDIO ENGINEERING SOCIETY, 2023, 71 (7-8): : 408 - 419
  • [7] Audio-Driven Talking Video Frame Restoration
    Cheng, Harry
    Guo, Yangyang
    Yin, Jianhua
    Chen, Haonan
    Wang, Jiafang
    Nie, Liqiang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 4110 - 4122
  • [8] Audio-Driven Facial Animation with Deep Learning: A Survey
    Jiang, Diqiong
    Chang, Jian
    You, Lihua
    Bian, Shaojun
    Kosk, Robert
    Maguire, Greg
    INFORMATION, 2024, 15 (11)
  • [9] Touch the Sound: Audio-Driven Tactile Feedback for Audio Mixing Applications
    Merchel, Sebastian
    Altinsoy, M. Ercan
    Stamm, Maik
    JOURNAL OF THE AUDIO ENGINEERING SOCIETY, 2012, 60 (1-2): : 47 - 53
  • [10] Multi-Task Audio-Driven Facial Animation
    Kim, Youngsoo
    An, Shounan
    Jo, Youngbak
    Park, Seungje
    Kang, Shindong
    Oh, Insoo
    Kim, Duke Donghyun
    SIGGRAPH '19 - ACM SIGGRAPH 2019 POSTERS, 2019,