Obstacle avoidance approach for quadruped robot based on multi-modal information fusion

被引:0
|
作者
Lyu, Youhao [1 ,2 ]
Jia, Yuanjun [2 ]
Zhuang, Yuan [2 ]
Dong, Qi [2 ]
机构
[1] Institute of Advanced Technology, University of Science and Technology of China, Hefei,230026, China
[2] China Academy of Electronics and Information Technology, Beijing,100049, China
关键词
Collision avoidance - Complex networks - Decision making - Motion planning - Multipurpose robots - Network architecture - Neural network models - Scattering parameters;
D O I
10.13374/j.issn2095-9389.2023.07.01.002
中图分类号
学科分类号
摘要
This paper proposes a multimodal information fusion neural network model that integrates visual, radar, and proprioceptive information. The model uses a spatial crossmodal attention mechanism to fuse the information, allowing the robot to focus on the most relevant information for obstacle avoidance. The attention mechanism enables the robot to selectively focus on the most relevant informative sensory inputs, which improves its ability to navigate complex terrain. The proposed method was evaluated using multiple experiments in challenging simulated environments, and the results showed a significant improvement in the obstacle avoidance success rate. The proposed method uses an actor–critic architecture and a proximal policy optimization (PPO) algorithm to train the robot in a simulated environment. The training process aims to reduce the difference between the robot’s performance in simulated and real-world environments. To achieve this, we randomly adjust the simulation environment’s parameters and add random noise to the robot’s sensory inputs. This approach allows the robot to learn a robust planning strategy that can be deployed in real-world environments. The multimodal information fusion neural network model is designed using a transformer-based architecture. The model shares the encoding of three types of tokens and generates features for the robot’s proprioceptive, visual, and point cloud inputs. The transformer encoder layers are stacked such that the token information from the three modalities can be fuzed at multiple levels. To balance the information from the three modalities, we first separately collect information for each modality and calculate the average value of all tokens from the same modality to obtain a single feature vector. This multimodal information fusion approach improves the robot’s decision-making capabilities in complex environments. The novelty of the proposed method lies in the introduction of a spatial crossmodal attention mechanism that allows the robot to selectively attend to the most informative sensory inputs. This attention mechanism improves the robot’s ability to navigate complex terrain and provides a certain degree of reliability for the quadruped robot in dynamic unknown environments. The combination of multimodal information fusion and attention mechanism enables the robot to adapt better to complex environments, thus improving its obstacle avoidance capabilities. Therefore, the proposed method provides a promising approach for improving the obstacle avoidance capabilities of quadruped robots in complex environments. The proposed method is based on the multimodal information fusion neural network model and spatial crossmodal attention mechanism. The experimental results demonstrate the effectiveness of the proposed method in improving the robot’s obstacle avoidance success rate. Moreover, the potential applications of the proposed method include search and rescue missions, exploration, and surveillance in complex environments. © 2024 Science Press. All rights reserved.
引用
收藏
页码:1426 / 1433
相关论文
共 50 条
  • [41] Multi-modal simultaneous machine translation fusion of image information
    Huang, Yan
    Wanga, Zhanyang
    Zhang, TianYuan
    Xu, Chun
    Lianga, Hui
    [J]. JOURNAL OF ENGINEERING RESEARCH, 2023, 11 (02):
  • [42] Multi-modal Information Extraction and Fusion with Convolutional Neural Networks
    Kumar, Dinesh
    Sharma, Dharmendra
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [43] Adaptive information fusion network for multi-modal personality recognition
    Bao, Yongtang
    Liu, Xiang
    Qi, Yue
    Liu, Ruijun
    Li, Haojie
    [J]. COMPUTER ANIMATION AND VIRTUAL WORLDS, 2024, 35 (03)
  • [44] Research on obstacle avoidance of a mobile robot based on visual information
    Jiang, Jin
    [J]. CIVIL, ARCHITECTURE AND ENVIRONMENTAL ENGINEERING, VOLS 1 AND 2, 2017, : 1479 - 1482
  • [45] Contextual Information Driven Multi-modal Medical Image Fusion
    Luo, Xiao-Qing
    Zhang, Zhan-Cheng
    Zhang, Bao-Cheng
    Wu, Xiao-Jun
    [J]. IETE TECHNICAL REVIEW, 2017, 34 (06) : 598 - 611
  • [46] Multi-modal Data Fusion for People Perception in the Social Robot Haru
    Ragel, Ricardo
    Rey, Rafael
    Paez, Lvaro
    Ponce, Javier
    Nakamura, Keisuke
    Caballero, Fernando
    Merino, Luis
    Gomez, Randy
    [J]. SOCIAL ROBOTICS, ICSR 2022, PT I, 2022, 13817 : 174 - 187
  • [47] Multi-Modal Sensor Fusion for Indoor Mobile Robot Pose Estimation
    Dobrev, Yassen
    Flores, Sergio
    Vossiek, Martin
    [J]. PROCEEDINGS OF THE 2016 IEEE/ION POSITION, LOCATION AND NAVIGATION SYMPOSIUM (PLANS), 2016, : 553 - 556
  • [48] A Sensor fusion technique using visual and ultrasonic information to acquire obstacle avoidance behaviors for quadruped robots
    Izumi, Kiyotaka
    Watanabe, Keigo
    Shindo, Masaaki
    Sato, Ryoichi
    [J]. 2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 3243 - +
  • [49] Analysis of a Wearable, Multi- modal Information Presentation Device for Obstacle Avoidance
    Gibson, Alison
    Webb, Andrea
    Stirling, Leia
    [J]. 2017 IEEE AEROSPACE CONFERENCE, 2017,
  • [50] Research on obstacle avoidance gait planning of quadruped crawling robot based on slope terrain recognition
    Wang, Peng
    Song, Chunxiao
    Dong, Renquan
    Zhang, Peng
    Yu, Shuang
    Zhang, Hao
    [J]. INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2022, 49 (05): : 1008 - 1021