Live Speech Driven Head-and-Eye Motion Generators

被引:50
|
作者
Le, Binh H. [1 ]
Ma, Xiaohan [1 ]
Deng, Zhigang
机构
[1] Univ Houston, Dept Comp Sci, Comp Graph Lab, Houston, TX 77204 USA
基金
美国国家科学基金会;
关键词
Facial animation; head and eye motion coupling; head motion synthesis; gaze synthesis; blinking model; live speech driven; ANIMATION; CAPTURE; MODEL; GAZE; PATTERNS; PROSODY; FACES;
D O I
10.1109/TVCG.2012.74
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
This paper describes a fully automated framework to generate realistic head motion, eye gaze, and eyelid motion simultaneously based on live (or recorded) speech input. Its central idea is to learn separate yet interrelated statistical models for each component (head motion, gaze, or eyelid motion) from a prerecorded facial motion data set: 1) Gaussian Mixture Models and gradient descent optimization algorithm are employed to generate head motion from speech features; 2) Nonlinear Dynamic Canonical Correlation Analysis model is used to synthesize eye gaze from head motion and speech features, and 3) nonnegative linear regression is used to model voluntary eye lid motion and log-normal distribution is used to describe involuntary eye blinks. Several user studies are conducted to evaluate the effectiveness of the proposed speech-driven head and eye motion generator using the well-established paired comparison methodology. Our evaluation results clearly show that this approach can significantly outperform the state-of-the-art head and eye motion generation algorithms. In addition, a novel mocap+video hybrid data acquisition technique is introduced to record high-fidelity head movement, eye gaze, and eyelid motion simultaneously.
引用
收藏
页码:1902 / 1914
页数:13
相关论文
共 50 条
  • [41] Head motion synthesis from speech using deep neural networks
    Ding, Chuang
    Xie, Lei
    Zhu, Pengcheng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2015, 74 (22) : 9871 - 9888
  • [42] Speech, Gaze and Head Motion in a Face-to-Face Collaborative Task
    Fagel, Sascha
    Bailly, Gerard
    TOWARD AUTONOMOUS, ADAPTIVE, AND CONTEXT-AWARE MULTIMODAL INTERFACES: THEORETICAL AND PRACTICAL ISSUES, 2011, 6456 : 256 - 264
  • [43] A control theoretic model of the combined planar motion of the human head and eye
    Egerstedt, M
    Martin, C
    APPLIED MATHEMATICS AND COMPUTATION, 1998, 90 (01) : 61 - 95
  • [44] Natural Eye Motion Synthesis by Modeling Gaze-Head Coupling
    Ma, Xiaohan
    Deng, Zhigang
    IEEE VIRTUAL REALITY 2009, PROCEEDINGS, 2009, : 143 - 150
  • [45] The attenuation of perceived motion smear during combined eye and head movements
    Tong, Jianliang
    Patel, Saumil S.
    Bedell, Harold E.
    VISION RESEARCH, 2006, 46 (26) : 4387 - 4397
  • [46] Rotational remapping in human spatial memory during eye and head motion
    Medendorp, WP
    Smith, MA
    Tweed, DB
    Crawford, JD
    JOURNAL OF NEUROSCIENCE, 2002, 22 (01):
  • [47] ITV CAMERA HEAD FOR TELEOPERATION ACTIVATED BY THE EYE MOTION OF THE OPERATOR.
    Iwata, Hiroo
    Ishii, Takemochi
    Hirose, Michitaka
    1600, (53):
  • [48] AFTEREFFECTS OF RETINAL MOTION RESULTING FROM HEAD OR EYE-MOVEMENTS
    SWANSTON, MT
    WADE, NJ
    PERCEPTION, 1988, 17 (03) : 362 - 363
  • [49] Blink-To-Live eye-based communication system for users with speech impairments
    Ezzat, Mohamed
    Maged, Mohamed
    Gamal, Youssef
    Adel, Mustafa
    Alrahmawy, Mohammed
    El-Metwally, Sara
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [50] Blink-To-Live eye-based communication system for users with speech impairments
    Mohamed Ezzat
    Mohamed Maged
    Youssef Gamal
    Mustafa Adel
    Mohammed Alrahmawy
    Sara El-Metwally
    Scientific Reports, 13