Learning to Shape by Grinding: Cutting-Surface-Aware Model-Based Reinforcement Learning

被引:2
|
作者
Hachimine, Takumi [1 ]
Morimoto, Jun [2 ,3 ]
Matsubara, Takamitsu [1 ]
机构
[1] Nara Inst Sci & Technol, Grad Sch Sci & Technol, Div Informat Sci, Nara 6300192, Japan
[2] Kyoto Univ, Grad Sch Informat, Dept Syst Sci, Kyoto 6068501, Japan
[3] Adv Telecommun Res Inst Int ATR, Brain Informat Commun Res Lab Grp BICR, Kyoto 6190288, Japan
关键词
Manipulation planning; model learning for control; reinforcement learning; OBJECTS;
D O I
10.1109/LRA.2023.3303721
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Object shaping by grinding is a crucial industrial process in which a rotating grinding belt removes material. Object-shape transition models are essential to achieving automation by robots; however, learning such a complex model that depends on process conditions is challenging because it requires a significant amount of data, and the irreversible nature of the removal process makes data collection expensive. This letter proposes a cutting-surface-aware Model-Based Reinforcement Learning (MBRL) method for robotic grinding. Our method employs a cutting-surface-aware model as the object's shape transition model, which in turn is composed of a geometric cutting model and a cutting-surface-deviation model, based on the assumption that the robot action can specify the cutting surface made by the tool. Furthermore, according to the grinding resistance theory, the cutting-surface-deviation model does not require raw shape information, making the model's dimensions smaller and easier to learn than a naive shape transition model directly mapping the shapes. Through evaluation and comparison by simulation and real robot experiments, we confirm that our MBRL method can achieve high data efficiency for learning object shaping by grinding and also provide generalization capability for initial and target shapes that differ from the training data.
引用
收藏
页码:6235 / 6242
页数:8
相关论文
共 50 条
  • [1] Value-Aware Loss Function for Model-based Reinforcement Learning
    Farahmand, Amir-massoud
    Barreto, Andre M. S.
    Nikovski, Daniel N.
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 54, 2017, 54 : 1486 - 1494
  • [2] Context-aware Dynamics Model for Generalization in Model-Based Reinforcement Learning
    Lee, Kimin
    Seo, Younggyo
    Lee, Seunghyun
    Lee, Honglak
    Shin, Jinwoo
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [3] A Safety Aware Model-Based Reinforcement Learning Framework for Systems with Uncertainties
    Mahmud, S. M. Nahid
    Hareland, Katrine
    Nivison, Scott A.
    Bell, Zachary, I
    Kamalapurkar, Rushikesh
    2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 1979 - 1984
  • [4] Delay-aware model-based reinforcement learning for continuous control
    Chen, Baiming
    Xu, Mengdi
    Li, Liang
    Zhao, Ding
    NEUROCOMPUTING, 2021, 450 : 119 - 128
  • [5] Learning to Paint With Model-based Deep Reinforcement Learning
    Huang, Zhewei
    Heng, Wen
    Zhou, Shuchang
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 8708 - 8717
  • [6] Model-based Reinforcement Learning: A Survey
    Moerland, Thomas M.
    Broekens, Joost
    Plaat, Aske
    Jonker, Catholijn M.
    FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2023, 16 (01): : 1 - 118
  • [7] A survey on model-based reinforcement learning
    Fan-Ming LUO
    Tian XU
    Hang LAI
    Xiong-Hui CHEN
    Weinan ZHANG
    Yang YU
    Science China(Information Sciences), 2024, 67 (02) : 59 - 84
  • [8] Nonparametric model-based reinforcement learning
    Atkeson, CG
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 1008 - 1014
  • [9] The ubiquity of model-based reinforcement learning
    Doll, Bradley B.
    Simon, Dylan A.
    Daw, Nathaniel D.
    CURRENT OPINION IN NEUROBIOLOGY, 2012, 22 (06) : 1075 - 1081
  • [10] Multiple model-based reinforcement learning
    Doya, K
    Samejima, K
    Katagiri, K
    Kawato, M
    NEURAL COMPUTATION, 2002, 14 (06) : 1347 - 1369