Integrated reinforcement and imitation learning for tower crane lift path planning

被引:4
|
作者
Wang, Zikang [1 ]
Huang, Chun [1 ]
Yao, Boqiang [2 ]
Li, Xin [3 ]
机构
[1] Beijing Univ Technol, Coll Architecture & Civil Engn, Beijing 100124, Peoples R China
[2] China Construct First Grp Construct & Dev Co Ltd, Beijing 100102, Peoples R China
[3] Educ Univ Hong Kong, Dept Math & Informat Technol, Hong Kong 999077, Peoples R China
关键词
Tower crane; Lift path planning; Reinforcement learning; Imitation learning; Virtual reality; VISUALIZATION; SIMULATION; SYSTEM;
D O I
10.1016/j.autcon.2024.105568
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Reinforcement learning (RL) has emerged as a promising solution method for crane-lift path planning. However, designing appropriate reward functions for tower crane (TC) operations remains particularly challenging. Poor design of reward functions can lead to non-executable lifting paths. This paper presents a framework combining imitation learning (IL) and RL to address the challenge. The framework comprises three steps: (1) designing a virtual environment consisting of construction site models and a TC model, (2) collecting expert demonstrations through virtual reality (VR) and pretraining through behavioral cloning (BC), and (3) refining the BC policies via generative adversarial imitation learning (GAIL) and proximal policy optimization (PPO). Using the paths generated by a PPO model as the baseline, the proposed BC + PPO + GAIL model exhibited better performance in both blind and nonblind lifting scenarios. This framework has been proven to generate realistic lifting paths mirroring crane operator behavior while ensuring efficiency and safety.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Reinforcement Learning for Altitude Hold and Path Planning in a Quadcopter
    Karthik, P. B.
    Kumar, Keshav
    Fernandes, Vikrant
    Arya, Kavi
    2020 6TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2020, : 463 - 467
  • [32] Path Planning for Autonomous Balloon Navigation with Reinforcement Learning
    He, Yingzhe
    Guo, Kai
    Wang, Chisheng
    Fu, Keyi
    Zheng, Jiehao
    ELECTRONICS, 2025, 14 (01):
  • [33] Path Planning for Construction Machines by Offline Reinforcement Learning
    Nakayama T.
    Kashi H.
    Uchimura Y.
    IEEJ Transactions on Industry Applications, 2024, 144 (05) : 367 - 373
  • [34] AGV Path Planning Model based on Reinforcement Learning
    Liao, Xiaofei
    Wang, Yang
    Xuan, Yiliang
    Wu, Dequan
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 6722 - 6726
  • [35] Reinforcement Learning Agent for Path Planning with Expert Demonstration
    Norkham, Alan
    Chalupa, Mikalus
    Gardner, Noah
    Khan, Md Abdullah Al Hafiz
    Zhang, Xinyue
    Hung, Chih-Cheng
    2022 IEEE 46TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2022), 2022, : 1016 - 1019
  • [36] Reinforcement Learning Path Planning Method with Error Estimation
    Zhang, Feihu
    Wang, Can
    Cheng, Chensheng
    Yang, Dianyu
    Pan, Guang
    ENERGIES, 2022, 15 (01)
  • [37] Robot path planning algorithm based on reinforcement learning
    Zhang F.
    Li N.
    Yuan R.
    Fu Y.
    Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2018, 46 (12): : 65 - 70
  • [38] Path planning of virtual human by using reinforcement learning
    He, Yue-Sheng
    Tang, Yuan-Yan
    PROCEEDINGS OF 2008 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2008, : 987 - 992
  • [39] Hierarchical Imitation and Reinforcement Learning
    Le, Hoang M.
    Jiang, Nan
    Agarwal, Alekh
    Dudik, Miroslav
    Yue, Yisong
    Daume, Hal, III
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [40] Delayed Reinforcement Learning by Imitation
    Liotet, Pierre
    Maran, Davide
    Bisi, Lorenzo
    Restelli, Marcello
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,