Text-Driven Video Prediction

被引:0
|
作者
Song, Xue [1 ]
Chen, Jingjing [1 ]
Zhu, Bin [2 ]
Jiang, Yu-Gang [1 ]
机构
[1] Shanghai Key Laboratory of Intelligent Information Processing, School of CS, Fudan University, Shanghai, China
[2] Singapore Management University, Singapore, Singapore
关键词
'current - Controllable video generation - Generation process - Motion inference - Motion information - Noise vectors - Stochastic generation - Text-driven video prediction - Video generation - Video prediction;
D O I
10.1145/3675171
中图分类号
学科分类号
摘要
Current video generation models usually convert signals indicating appearance and motion received from inputs (e.g., image and text) or latent spaces (e.g., noise vectors) into consecutive frames, fulfilling a stochastic generation process for the uncertainty introduced by latent code sampling. However, this generation pattern lacks deterministic constraints for both appearance and motion, leading to uncontrollable and undesirable outcomes. To this end, we propose a new task called Text-driven Video Prediction (TVP). Taking the first frame and text caption as inputs, this task aims to synthesize the following frames. Specifically, appearance and motion components are provided by the image and caption separately. The key to addressing the TVP task depends on fully exploring the underlying motion information in text descriptions, thus facilitating plausible video generation. In fact, this task is intrinsically a cause-and-effect problem, as the text content directly influences the motion changes of frames. To investigate the capability of text in causal inference for progressive motion information, our TVP framework contains a Text Inference Module (TIM), producing stepwise embeddings to regulate motion inference for subsequent frames. In particular, a refinement mechanism incorporating global motion semantics guarantees coherent generation. Extensive experiments are conducted on Something-Something V2 and Single Moving MNIST datasets. Experimental results demonstrate that our model achieves better results over other baselines, verifying the effectiveness of the proposed framework. © 2024 Copyright held by the owner/author(s)
引用
收藏
相关论文
共 50 条
  • [1] Text2Performer: Text-Driven Human Video Generation
    Jiang, Yuming
    Yang, Shuai
    Koh, Tong Liang
    Wu, Wayne
    Loy, Chen Change
    Liu, Ziwei
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 22690 - 22700
  • [2] Shape-aware Text-driven Layered Video Editing
    Lee, Yao-Chih
    Jang, Ji-Ze Genevieve
    Chen, Yi-Ting
    Qiu, Elizabeth
    Huang, Jia-Bin
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 14317 - 14326
  • [3] Explainable Text-Driven Neural Network for Stock Prediction
    Yang, Linyi
    Zhang, Zheng
    Xiong, Su
    Wei, Lirui
    Ng, James
    Xu, Lina
    Dong, Ruihai
    [J]. PROCEEDINGS OF 2018 5TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2018, : 441 - 445
  • [4] Text2LIVE: Text-Driven Layered Image and Video Editing
    Bar-Tal, Omer
    Ofri-Amar, Dolev
    Fridman, Rafail
    Kasten, Yoni
    Dekel, Tali
    [J]. COMPUTER VISION - ECCV 2022, PT XV, 2022, 13675 : 707 - 723
  • [5] StableVideo: Text-driven Consistency-aware Diffusion Video Editing
    Chai, Wenhao
    Guo, Xun
    Wang, Gaoang
    Lu, Yan
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 22983 - 22993
  • [6] Text2Video: Text-driven facial animation using MPEG-4
    Rurainsky, J
    Eisert, P
    [J]. VISUAL COMMUNICATIONS AND IMAGE PROCESSING 2005, PTS 1-4, 2005, 5960 : 492 - 500
  • [7] A statistical parametric approach to video-realistic text-driven talking avatar
    Lei Xie
    Naicai Sun
    Bo Fan
    [J]. Multimedia Tools and Applications, 2014, 73 : 377 - 396
  • [8] TEXT2VIDEO: TEXT-DRIVEN TALKING-HEAD VIDEO SYNTHESIS WITH PERSONALIZED PHONEME - POSE DICTIONARY
    Zhang, Sibo
    Yuan, Jiahong
    Liao, Miao
    Zhang, Liangjun
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2659 - 2663
  • [9] A statistical parametric approach to video-realistic text-driven talking avatar
    Xie, Lei
    Sun, Naicai
    Fan, Bo
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2014, 73 (01) : 377 - 396
  • [10] Hybrid Deep Sequential Modeling for Social Text-Driven Stock Prediction
    Wu, Huizhe
    Zhang, Wei
    Shen, Weiwei
    Wang, Jun
    [J]. CIKM'18: PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2018, : 1627 - 1630