Imitation Learning from a Single Demonstration Leveraging Vector Quantization for Robotic Harvesting

被引:0
|
作者
Porichis, Antonios [1 ,2 ]
Inglezou, Myrto [1 ]
Kegkeroglou, Nikolaos [3 ]
Mohan, Vishwanathan [1 ]
Chatzakos, Panagiotis [1 ]
机构
[1] Univ Essex, AI Innovat Ctr, Wivenhoe Pk, Colchester CO4 3SQ, England
[2] Natl Struct Integr Res Ctr, Granta Pk, Cambridge CB21 6AL, England
[3] TWI Hellas, 280 Kifisias Ave, Halandri 15232, Greece
基金
欧盟地平线“2020”;
关键词
imitation learning; learning by demonstration; vector quantization; mushroom harvesting; visual servoing;
D O I
10.3390/robotics13070098
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
The ability of robots to tackle complex non-repetitive tasks will be key in bringing a new level of automation in agricultural applications still involving labor-intensive, menial, and physically demanding activities due to high cognitive requirements. Harvesting is one such example as it requires a combination of motions which can generally be broken down into a visual servoing and a manipulation phase, with the latter often being straightforward to pre-program. In this work, we focus on the task of fresh mushroom harvesting which is still conducted manually by human pickers due to its high complexity. A key challenge is to enable harvesting with low-cost hardware and mechanical systems, such as soft grippers which present additional challenges compared to their rigid counterparts. We devise an Imitation Learning model pipeline utilizing Vector Quantization to learn quantized embeddings directly from visual inputs. We test this approach in a realistic environment designed based on recordings of human experts harvesting real mushrooms. Our models can control a cartesian robot with a soft, pneumatically actuated gripper to successfully replicate the mushroom outrooting sequence. We achieve 100% success in picking mushrooms among distractors with less than 20 min of data collection comprising a single expert demonstration and auxiliary, non-expert, trajectories. The entire model pipeline requires less than 40 min of training on a single A4000 GPU and approx. 20 ms for inference on a standard laptop GPU.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Visual Imitation Learning for robotic fresh mushroom harvesting
    Porichis, Antonios
    Vasios, Konstantinos
    Iglezou, Myrto
    Mohan, Vishwanathan
    Chatzakos, Panagiotis
    2023 31ST MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION, MED, 2023, : 535 - 540
  • [2] Leveraging Imitation Learning on Pose Regulation Problem of a Robotic Fish
    Zhang, Tianhao
    Yue, Lu
    Wang, Chen
    Sun, Jinan
    Zhang, Shikun
    Wei, Airong
    Xie, Guangming
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 4232 - 4245
  • [3] Generative Adversarial Network for Imitation Learning from Single Demonstration
    Tho Nguyen Duc
    Chanh Minh Tran
    Phan Xuan Tan
    Kamioka, Eiji
    BAGHDAD SCIENCE JOURNAL, 2021, 18 (04) : 1350 - 1355
  • [4] Imitation Learning from Imperfect Demonstration
    Wu, Yueh-Hua
    Charoenphakdee, Nontawat
    Bao, Han
    Tangkaratt, Voot
    Sugiyama, Masashi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] Imitation Learning from Video by Leveraging Proprioception
    Torabi, Faraz
    Warnell, Garrett
    Stone, Peter
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3585 - 3591
  • [6] A Computational Framework for Integrating Robotic Exploration and Human Demonstration in Imitation Learning
    Tan, Huan
    Kawamura, Kazuhiko
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 2501 - 2506
  • [7] Coarse-to-Fine Imitation Learning: Robot Manipulation from a Single Demonstration
    Johns, Edward
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4613 - 4619
  • [8] Goal-aware generative adversarial imitation learning from imperfect demonstration for robotic cloth manipulation
    Tsurumine, Yoshihisa
    Matsubara, Takamitsu
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2022, 158
  • [9] Leveraging Temporal Reasoning for Policy Selection in Learning from Demonstration
    Carpio, Estuardo
    Clark-Turner, Madison
    Gesel, Paul
    Begum, Momotaz
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 7798 - 7804
  • [10] On the Guaranteed Almost Equivalence Between Imitation Learning From Observation and Demonstration
    Cheng, Zhihao
    Liu, Liu
    Liu, Aishan
    Sun, Hao
    Fang, Meng
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) : 677 - 689