Multi-pedestrian trajectory prediction method based on multi-view 3D simulation video learning

被引:0
|
作者
Cao, Xingwen [1 ,2 ]
Zheng, Hongwei [1 ,2 ]
Liu, Ying [1 ,2 ]
Wu, Mengquan [3 ]
Wang, Lingyue [1 ,2 ]
Bao, Anming [1 ,2 ]
Chen, Xi [1 ,2 ]
机构
[1] State Key Laboratory of Desert and Oasis Ecology, Xinjiang Institute of Ecology and Geography, Chinese Academy of Sciences, Urumqi,830011, China
[2] College of Resources and Environment, University of Chinese Academy of Sciences, Beijing,100049, China
[3] School of Resources and Environmental Engineering, Ludong University, Yantai,264025, China
基金
中国国家自然科学基金;
关键词
Decoding - Forecasting - Geographic information systems - Information systems - Information use - Learning systems - Recurrent neural networks - Urban transportation;
D O I
10.11947/j.AGCS.2023.20220239
中图分类号
学科分类号
摘要
Multi-pedestrian trajectory prediction is one of the key factors in integrating urban geographic information system and intelligent transportation. To address the problems of insufficient training data, difficult labeling, and low accuracy of pedestrian trajectory prediction in multi-view scenes for existing methods, we propose a novel multi-pedestrian trajectory prediction method based on multi-view 3D simulation video learning. First, a simulation simulator is used to generate the required multi-view pedestrian trajectory annotation data. Then, we mix up the trajectory of the selected view and the adversarial trajectory by a convex combination function to generate the enhanced adversarial trajectory. Next, an advanced detection and tracking algorithm is used to encode and track pedestrian appearance information. Furthermore, the enhanced trajectory and coding information are used as the feature input of a graph attention recurrent neural network to model pedestrian interaction. Finally, the pedestrian trajectory is decoded by a position decoder to extract pedestrian motion characteristics, and multi-pedestrian trajectory prediction is completed. The ADE and FDE accuracies of our method on the ETH/UCY fixed-view dataset are 0.41 and 0.82, respectively. The ADE accuracy on the ActEV/VIRAT and Argoverse multi-view datasets is 17.74 and 65.4, and the FDE accuracy is 34.96 and 172.8. © 2023 SinoMaps Press. All rights reserved.
引用
收藏
页码:1595 / 1608
相关论文
共 50 条
  • [31] Overview of 3D Reconstruction Methods Based on Multi-view
    Li, Mengxin
    Zheng, Dai
    Zhang, Rui
    Yin, Jiadi
    Tian, Xiangqian
    [J]. 2015 7TH INTERNATIONAL CONFERENCE ON INTELLIGENT HUMAN-MACHINE SYSTEMS AND CYBERNETICS IHMSC 2015, VOL II, 2015,
  • [32] Underwater 3D reconstruction based on multi-view stereo
    Gu, Feifei
    Zhao, Juan
    Xu, Pei
    Huang, Shulan
    Zhang, Gaopeng
    Song, Zhan
    [J]. OCEAN OPTICS AND INFORMATION TECHNOLOGY, 2018, 10850
  • [33] MULTI-VIEW PAIRWISE RELATIONSHIP LEARNING FOR SKETCH BASED 3D SHAPE RETRIEVAL
    Li, Hanhui
    Wu, Hefeng
    He, Xiangjian
    Lin, Shujin
    Wang, Ruomei
    Luo, Xiaonan
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 1434 - 1439
  • [34] Multi-view 3D reconstruction based on deep learning: A survey and comparison of methods
    Wu, Juhao
    Wyman, Omar
    Tang, Yadong
    Pasini, Damiano
    Wang, Wenlong
    [J]. Neurocomputing, 2024, 582
  • [35] 3D Reconstruction for Multi-view Objects
    Yu, Jun
    Yin, Wenbin
    Hu, Zhiyi
    Liu, Yabin
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2023, 106
  • [36] Multi-view 3D reconstruction based on deep learning: A survey and comparison of methods
    Wu, Juhao
    Wyman, Omar
    Tang, Yadong
    Pasini, Damiano
    Wang, Wenlong
    [J]. NEUROCOMPUTING, 2024, 582
  • [37] Review of multi-view 3D object recognition methods based on deep learning
    Qi, Shaohua
    Ning, Xin
    Yang, Guowei
    Zhang, Liping
    Long, Peng
    Cai, Weiwei
    Li, Weijun
    [J]. DISPLAYS, 2021, 69
  • [38] Multi-view 3D Reconstruction with Transformers
    Wang, Dan
    Cui, Xinrui
    Chen, Xun
    Zou, Zhengxia
    Shi, Tianyang
    Salcudean, Septimiu
    Wang, Z. Jane
    Ward, Rabab
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 5702 - 5711
  • [39] 3D-DCT Chip Design for 3D Multi-view Video Compression
    Fan, Yu-Cheng
    Chen, Shan-Ann
    Wu, Kuo-Gi
    You, Jun-Lin
    [J]. APPLIED MATHEMATICS & INFORMATION SCIENCES, 2012, 6 (02): : 567S - 572S
  • [40] Wide-Baseline Multi-View Video Segmentation for 3D Reconstruction
    Sarim, Muhammad
    Hilton, Adrian
    Guillemaut, Jean-Yves
    Kim, Hansung
    Takai, Takeshi
    [J]. PROCEEDINGS OF THE 2010 ACM WORKSHOP ON 3D VIDEO PROCESSING (3DVP'10), 2010, : 13 - 18