Obstacle avoidance approach for quadruped robot based on multi-modal information fusion

被引:0
|
作者
Lyu, Youhao [1 ,2 ]
Jia, Yuanjun [2 ]
Zhuang, Yuan [2 ]
Dong, Qi [2 ]
机构
[1] Institute of Advanced Technology, University of Science and Technology of China, Hefei,230026, China
[2] China Academy of Electronics and Information Technology, Beijing,100049, China
关键词
Reinforcement learning;
D O I
10.13374/j.issn2095-9389.2023.07.01.002
中图分类号
学科分类号
摘要
This paper proposes a multimodal information fusion neural network model that integrates visual, radar, and proprioceptive information. The model uses a spatial crossmodal attention mechanism to fuse the information, allowing the robot to focus on the most relevant information for obstacle avoidance. The attention mechanism enables the robot to selectively focus on the most relevant informative sensory inputs, which improves its ability to navigate complex terrain. The proposed method was evaluated using multiple experiments in challenging simulated environments, and the results showed a significant improvement in the obstacle avoidance success rate. The proposed method uses an actor–critic architecture and a proximal policy optimization (PPO) algorithm to train the robot in a simulated environment. The training process aims to reduce the difference between the robot’s performance in simulated and real-world environments. To achieve this, we randomly adjust the simulation environment’s parameters and add random noise to the robot’s sensory inputs. This approach allows the robot to learn a robust planning strategy that can be deployed in real-world environments. The multimodal information fusion neural network model is designed using a transformer-based architecture. The model shares the encoding of three types of tokens and generates features for the robot’s proprioceptive, visual, and point cloud inputs. The transformer encoder layers are stacked such that the token information from the three modalities can be fuzed at multiple levels. To balance the information from the three modalities, we first separately collect information for each modality and calculate the average value of all tokens from the same modality to obtain a single feature vector. This multimodal information fusion approach improves the robot’s decision-making capabilities in complex environments. The novelty of the proposed method lies in the introduction of a spatial crossmodal attention mechanism that allows the robot to selectively attend to the most informative sensory inputs. This attention mechanism improves the robot’s ability to navigate complex terrain and provides a certain degree of reliability for the quadruped robot in dynamic unknown environments. The combination of multimodal information fusion and attention mechanism enables the robot to adapt better to complex environments, thus improving its obstacle avoidance capabilities. Therefore, the proposed method provides a promising approach for improving the obstacle avoidance capabilities of quadruped robots in complex environments. The proposed method is based on the multimodal information fusion neural network model and spatial crossmodal attention mechanism. The experimental results demonstrate the effectiveness of the proposed method in improving the robot’s obstacle avoidance success rate. Moreover, the potential applications of the proposed method include search and rescue missions, exploration, and surveillance in complex environments. © 2024 Science Press. All rights reserved.
引用
收藏
页码:1426 / 1433
相关论文
共 50 条
  • [1] Obstacle avoidance technology of bionic quadruped robot based on multi-sensor information fusion
    韩宝玲
    张天
    罗庆生
    朱颖
    宋明辉
    [J]. Journal of Beijing Institute of Technology, 2016, 25 (04) : 448 - 454
  • [2] Mobile Robot Obstacle Avoidance Based on Multi-sensor Information Fusion Technology
    Cheng, Yuanhang
    Zhang, Chunlan
    [J]. MECHANICAL DESIGN AND POWER ENGINEERING, PTS 1 AND 2, 2014, 490-491 : 1168 - 1171
  • [3] Likelihood Confidence Rating Based Multi-Modal Information Fusion for Robot Fine Operation
    Xiao, Wei
    Liu, Hong
    Sun, Fuchun
    Liu, Huaping
    [J]. 2014 13TH INTERNATIONAL CONFERENCE ON CONTROL AUTOMATION ROBOTICS & VISION (ICARCV), 2014, : 265 - 270
  • [4] Robot obstacle avoidance and navigation control research based on multi-sensor information fusion
    Fan, Xiaojing
    Jiang, Mingyang
    Pei, Zhili
    [J]. BASIC & CLINICAL PHARMACOLOGY & TOXICOLOGY, 2018, 124 : 51 - 52
  • [5] ALPHRED: A Multi-Modal Operations Quadruped Robot for Package Delivery Applications
    Hooks, Joshua
    Ahn, Min Sung
    Yu, Jeffrey
    Zhang, Xiaoguang
    Zhu, Taoyuanmin
    Chae, Hosik
    Hong, Dennis
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04): : 5409 - 5416
  • [6] Obstacle Avoidance Research of Snake-like Robot based on Multi-sensor Information Fusion
    Wu, Qianying
    Gao, Junyao
    Huang, Chengzu
    Zhao, Zhengyang
    Wang, Cheng
    Su, Xuandong
    Liu, Huaxin
    Li, Xin
    Liu, Yi
    Xu, Zhe
    [J]. 2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO 2012), 2012,
  • [7] Predict Robot Grasp Outcomes based on Multi-Modal Information
    Yang, Chao
    Du, Peng
    Sun, Fuchun
    Fang, Bin
    Zhou, Jie
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2018, : 1563 - 1568
  • [8] Robot Obstacle Avoidance and Navigation Control Algorithm Research Based on Multi-sensor Information Fusion
    Wang, Ziyang
    [J]. 2018 11TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTATION TECHNOLOGY AND AUTOMATION (ICICTA 2018), 2018, : 351 - 354
  • [9] Visual Sorting Method Based on Multi-Modal Information Fusion
    Han, Song
    Liu, Xiaoping
    Wang, Gang
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (06):
  • [10] News video classification based on multi-modal information fusion
    Lie, WN
    Su, CK
    [J]. 2005 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), VOLS 1-5, 2005, : 1021 - 1024