Attention-Based Bidirectional Recurrent Neural Networks for Description Generation of Videos

被引:0
|
作者
Du, Xiaotong [1 ]
Yuan, Jiabin [1 ]
Liu, Hu [1 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing, Jiangsu, Peoples R China
来源
CLOUD COMPUTING AND SECURITY, PT VI | 2018年 / 11068卷
关键词
Video description; Convolutional Neural Networks; Bidirectional Recurrent Neural Networks; Attention mechanism;
D O I
10.1007/978-3-030-00021-9_40
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Describing videos in human language is of vital importance in many applications, such as managing massive videos on line and providing descriptive video service (DVS) for blind people. In order to further promote existing video description frameworks, this paper presents an end-to-end deep learning model incorporating Convolutional Neural Networks (CNNs) and Bidirectional Recurrent Neural Networks (BiRNNs) based on a multimodal attention mechanism. Firstly, the model produces richer video representations, including image feature, motion feature and audio feature, than other similar researches. Secondly, BiRNNs model encodes these features in both forward and backward directions. Finally, an attention-based decoder translates sequential outputs of encoder to sequential words. The model is evaluated on Microsoft Research Video Description Corpus (MSVD) dataset. The results demonstrate the necessity of combining BiRNNs with a multimodal attention mechanism and the superiority of this model over other state-of-the-art methods conducted on this dataset.
引用
收藏
页码:440 / 451
页数:12
相关论文
共 50 条
  • [41] Attention-based recurrent neural network for influenza epidemic prediction
    Zhu, Xianglei
    Fu, Bofeng
    Yang, Yaodong
    Ma, Yu
    Hao, Jianye
    Chen, Siqi
    Liu, Shuang
    Li, Tiegang
    Liu, Sen
    Guo, Weiming
    Liao, Zhenyu
    BMC BIOINFORMATICS, 2019, 20 (Suppl 18)
  • [42] Seizure localisation with attention-based graph neural networks
    Grattarola, Daniele
    Livi, Lorenzo
    Alippi, Cesare
    Wennberg, Richard
    Valiante, Taufik A.
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 203
  • [43] Causal Discovery with Attention-Based Convolutional Neural Networks
    Nauta, Meike
    Bucur, Doina
    Seifert, Christin
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2019, 1 (01):
  • [44] MPA-RNN: A Novel Attention-Based Recurrent Neural Networks for Total Nitrogen Prediction
    Geng, Jingxuan
    Yang, Chunhua
    Li, Yonggang
    Lan, Lijuan
    Luo, Qiwu
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (10) : 6516 - 6525
  • [45] VIDEOWHISPER: Toward Discriminative Unsupervised Video Feature Learning With Attention-Based Recurrent Neural Networks
    Zhao, Na
    Zhang, Hanwang
    Hong, Richang
    Wang, Meng
    Chua, Tat-Seng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2017, 19 (09) : 2080 - 2092
  • [46] Attention-based recurrent neural network for influenza epidemic prediction
    Xianglei Zhu
    Bofeng Fu
    Yaodong Yang
    Yu Ma
    Jianye Hao
    Siqi Chen
    Shuang Liu
    Tiegang Li
    Sen Liu
    Weiming Guo
    Zhenyu Liao
    BMC Bioinformatics, 20
  • [47] Cascade Dynamics Modeling with Attention-based Recurrent Neural Network
    Wang, Yongqing
    Shen, Huawei
    Liu, Shenghua
    Gao, Jinhua
    Cheng, Xueqi
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2985 - 2991
  • [48] Attention-Based Recurrent Neural Network for Plant Disease Classification
    Lee, Sue Han
    Goeau, Herve
    Bonnet, Pierre
    Joly, Alexis
    FRONTIERS IN PLANT SCIENCE, 2020, 11
  • [49] Signal Peptides Generated by Attention-Based Neural Networks
    Wu, Zachary
    Yang, Kevin K.
    Liszka, Michael J.
    Lee, Alycia
    Batzilla, Alina
    Wernick, David
    Weiner, David P.
    Arnold, Frances H.
    ACS SYNTHETIC BIOLOGY, 2020, 9 (08): : 2154 - 2161
  • [50] Attention-Based Hierarchical Recurrent Neural Network for Phenotype Classification
    Xu, Nan
    Shen, Yanyan
    Zhu, Yanmin
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2019, PT I, 2019, 11439 : 465 - 476