Automatic report generation based on multi-modal information

被引:0
|
作者
Jing Zhang
Xiaoxue Li
Weizhi Nie
Yuting Su
机构
[1] Tianjin University,School of Electronics Information Engineering
来源
关键词
News event detection; Multi-modal; Report generation;
D O I
暂无
中图分类号
学科分类号
摘要
In this paper, we propose a new framework which can utilize multi-modal social media information to automatically generate related reports for users or government. First, we utilize DBSCAN (Density Based Spatial Clustering of Applications with Noise) to detect events in official news websites. Then, some unofficial information details are extracted from social network platforms (Foursquare, Twitter, YouTube), which will be leveraged to enhance the official report in order to excavate some latent and useful information. In this process, we applied some classic textual processing methods and computer vision technologies to reduce the noise information uploaded by user generated contents (UGCs). Then, we applied LSTM-CNN model to generate the related image caption and successfully convert visual information to textual information. Finally, we extracted some latent topics using graph cluster method to generate the final report. To demonstrate the effectiveness of our framework, we got a large of multi-source event dataset from official news websites and Twitter. Finally, the user study demonstrates the practicability of our approach.
引用
收藏
页码:12005 / 12015
页数:10
相关论文
共 50 条
  • [1] Automatic report generation based on multi-modal information
    Zhang, Jing
    Li, Xiaoxue
    Nie, Weizhi
    Su, Yuting
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (09) : 12005 - 12015
  • [2] Automatic Medical Image Report Generation with Multi-view and Multi-modal Attention Mechanism
    Yang, Shaokang
    Niu, Jianwei
    Wu, Jiyan
    Liu, Xuefeng
    [J]. ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT III, 2020, 12454 : 687 - 699
  • [3] Automatic Inspection of Railway Carbon Strips Based on Multi-Modal Visual Information
    Di Stefano, Erika
    Avizzano, Carlo Alberto
    Bergamasco, Massimo
    Masini, Paolo
    Menci, Mauro
    Russo, Davide
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2017, : 178 - 184
  • [4] Automatic generation of multi-modal dialogue from text based on discourse structure analysis
    Prendinger, Helmut
    Piwek, Paul
    Ishizuka, Mitsuru
    [J]. ICSC 2007: INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING, PROCEEDINGS, 2007, : 27 - +
  • [5] Multi-modal Fusion Based Automatic Pain Assessment
    Zhi, Ruicong
    Yu, Junwei
    [J]. PROCEEDINGS OF 2019 IEEE 8TH JOINT INTERNATIONAL INFORMATION TECHNOLOGY AND ARTIFICIAL INTELLIGENCE CONFERENCE (ITAIC 2019), 2019, : 1378 - 1382
  • [6] Lightweight multi-modal emotion recognition model based on modal generation
    Liu, Peisong
    Che, Manqiang
    Luo, Jiangchuan
    [J]. 2022 9TH INTERNATIONAL FORUM ON ELECTRICAL ENGINEERING AND AUTOMATION, IFEEA, 2022, : 430 - 435
  • [7] MATNet: Exploiting Multi-Modal Features for Radiology Report Generation
    Shang, Caozhi
    Cui, Shaoguo
    Li, Tiansong
    Wang, Xi
    Li, Yongmei
    Jiang, Jingfeng
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2692 - 2696
  • [8] Multi-modal visual tracking based on textual generation
    Wang, Jiahao
    Liu, Fang
    Jiao, Licheng
    Wang, Hao
    Li, Shuo
    Li, Lingling
    Chen, Puhua
    Liu, Xu
    [J]. INFORMATION FUSION, 2024, 112
  • [9] Synthetic Sensor Measurement Generation With Noise Learning and Multi-Modal Information
    Romanelli, Fabrizio
    Martinelli, Francesco
    [J]. IEEE ACCESS, 2023, 11 : 111765 - 111788
  • [10] Automatic controller generation based on dependency network of multi-modal sensor variables for musculoskeletal robotic arm
    Kobayashi, Yuichi
    Harada, Kentaro
    Takagi, Kentaro
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 118 : 55 - 65