A Vision-Based Bio-Inspired Reinforcement Learning Algorithms for Manipulator Obstacle Avoidance

被引:1
|
作者
Singh, Abhilasha [1 ]
Shakeel, Mohamed [2 ]
Kalaichelvi, V [1 ]
Karthikeyan, R. [2 ]
机构
[1] Birla Inst Technol & Sci Pilani, Dept Elect & Elect Engn, Dubai Campus,POB 345 055, Dubai, U Arab Emirates
[2] Birla Inst Technol & Sci Pilani, Dept Mech Engn, Dubai Campus,POB 345 055, Dubai, U Arab Emirates
关键词
Q-learning; DQN; SARSA; DDQN; homogeneous transformation; optimization; obstacle avoidance; MOBILE ROBOT; ENVIRONMENTS;
D O I
10.3390/electronics11213636
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Path planning for robotic manipulators has proven to be a challenging issue in industrial applications. Despite providing precise waypoints, the traditional path planning algorithm requires a predefined map and is ineffective in complex, unknown environments. Reinforcement learning techniques can be used in cases where there is a no environmental map. For vision-based path planning and obstacle avoidance in assembly line operations, this study introduces various Reinforcement Learning (RL) algorithms based on discrete state-action space, such as Q-Learning, Deep Q Network (DQN), State-Action-Reward-State-Action (SARSA), and Double Deep Q Network (DDQN). By positioning the camera in an eye-to-hand position, this work used color-based segmentation to identify the locations of obstacles, start, and goal points. The homogeneous transformation technique was used to further convert the pixel values into robot coordinates. Furthermore, by adjusting the number of episodes, steps per episode, learning rate, and discount factor, a performance study of several RL algorithms was carried out. To further tune the training hyperparameters, genetic algorithms (GA) and particle swarm optimization (PSO) were employed. The length of the path travelled, the average reward, the average number of steps, and the time required to reach the objective point were all measured and compared for each of the test cases. Finally, the suggested methodology was evaluated using a live camera that recorded the robot workspace in real-time. The ideal path was then drawn using a TAL BRABO 5 DOF manipulator. It was concluded that waypoints obtained via Double DQN showed an improved performance and were able to avoid the obstacles and reach the goal point smoothly and efficiently.
引用
收藏
页数:26
相关论文
共 50 条
  • [21] Bio-Inspired Model for Gestures Recognition through Vision-based Movement Primitives
    Nope, Sandra E.
    Loaiza, Humberto
    Caicedo, Eduardo
    REVISTA IBEROAMERICANA DE AUTOMATICA E INFORMATICA INDUSTRIAL, 2008, 5 (04): : 69 - +
  • [22] Bio-Inspired Cooperative Control Scheme of Obstacle Avoidance for UUV Swarm
    Wang, Zhao
    Wang, Hongjian
    Yuan, Jianya
    Yu, Dan
    Zhang, Kai
    Ren, Jingfei
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2024, 12 (03)
  • [23] Bio-inspired vision
    Posch, C.
    JOURNAL OF INSTRUMENTATION, 2012, 7
  • [24] Vision-based Obstacle Avoidance Algorithm for Mobile Robot
    Li, Yong
    Liu, Yuzhe
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1273 - 1278
  • [25] Vision-Based Obstacle Avoidance Using SIFT Features
    Chavez, Aaron
    Gustafson, David
    ADVANCES IN VISUAL COMPUTING, PT 2, PROCEEDINGS, 2009, 5876 : 550 - 557
  • [26] Uncalibrated vision-based mobile robot obstacle avoidance
    Piepmeier, JA
    PROCEEDINGS OF THE 33RD SOUTHEASTERN SYMPOSIUM ON SYSTEM THEORY, 2001, : 251 - 255
  • [27] Vision-based UAV flight control and obstacle avoidance
    He, Zhihai
    Venkataraman Iyer, Ram
    Chandler, Phillip R.
    2006 AMERICAN CONTROL CONFERENCE, VOLS 1-12, 2006, 1-12 : 2166 - +
  • [28] Monocular Vision-Based Obstacle Detection and Avoidance for a Multicopter
    Chen, Hsiang-Chieh
    IEEE ACCESS, 2019, 7 : 167869 - 167883
  • [29] Bio-Inspired Vision-Based Leader-Follower Formation Flying in the Presence of Delays
    Oyekan, John
    ROBOTICS, 2016, 5 (03)
  • [30] Embedded and real-time architecture for bio-inspired vision-based robot navigation
    Laurent Fiack
    Nicolas Cuperlier
    Benoît Miramond
    Journal of Real-Time Image Processing, 2015, 10 : 699 - 722