Dynamic assembly sequence selection using reinforcement learning

被引:9
|
作者
Lowe, G [1 ]
Shirinzadeh, B [1 ]
机构
[1] Monash Univ, Sch Comp Sci & Software Engn, Clayton, Vic 3168, Australia
关键词
D O I
10.1109/ROBOT.2004.1307458
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Determining the most appropriate sequence for assembling products requires assessment of the process, product, and the technology applied. Most production engineers apply constraint based evaluation and history to identify the solution sequence. What if their solution is sub-optimal? In this paper a self-learning technique for selecting a sequence and dynamically changing the sequence is presented, selection is based on the history of assemblies. The evaluation is dependent on part properties rather than parts and their relationships, thus no previous knowledge of parts and their interaction is required in the decision making process. The method assumes assembly is without constraint, for example, a highly flexible robotic assembly cell. This maximises the ability of the algorithm to select sequences for new products and optimse them. The heart of the algorithm is a reinforcement learning model which punishes failed assembly steps, this facilitates feedback sequence selection, where current methods are merely feedforward. This feedback approach addresses combinatorial explosion that can cripple assembly planners.
引用
收藏
页码:2633 / 2638
页数:6
相关论文
共 50 条
  • [1] Dynamic Algorithm Selection Using Reinforcement Learning
    Armstrong, Warren
    Christen, Peter
    McCreath, Eric
    Rendell, Alistair P.
    [J]. AIDM 2006: INTERNATIONAL WORKSHOP ON INTEGRATING AI AND DATING MINING, 2006, : 18 - +
  • [2] ASSEMBLY SEQUENCE OPTIMIZATION OF SPATIAL TRUSSES USING GRAPH EMBEDDING AND REINFORCEMENT LEARNING
    Hayashi, Kazuki
    Ohsaki, Makoto
    Kotera, Masaya
    [J]. JOURNAL OF THE INTERNATIONAL ASSOCIATION FOR SHELL AND SPATIAL STRUCTURES, 2022, 63 (04): : 232 - 240
  • [3] Assembly sequence planning based on deep reinforcement learning
    Zhao M.-H.
    Zhang X.-B.
    Guo X.
    Ou Y.-S.
    [J]. Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2021, 38 (12): : 1901 - 1910
  • [4] A reinforcement learning approach for dynamic supplier selection
    Kim, Tae Il
    Bilsel, R. Ufuk
    Kumara, Soundar R. T.
    [J]. PROCEEDINGS OF THE 2007 IEEE INTERNATIONAL CONFERENCE ON SERVICE OPERATIONS AND LOGISTICS, AND INFORMATICS, 2007, : 19 - +
  • [5] Genome Assembly Using Reinforcement Learning
    Xavier, Roberto
    de Souza, Kleber Padovani
    Chateau, Annie
    Alves, Ronnie
    [J]. ADVANCES IN BIOINFORMATICS AND COMPUTATIONAL BIOLOGY, BSB 2019, 2020, 11347 : 16 - 28
  • [6] Reinforcement learning of dynamic motor sequence: Learning to stand up
    Morimoto, J
    Doya, K
    [J]. 1998 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS - PROCEEDINGS, VOLS 1-3: INNOVATIONS IN THEORY, PRACTICE AND APPLICATIONS, 1998, : 1721 - 1726
  • [7] FINDING THE OPTIMAL SEQUENCE OF FEATURES SELECTION BASED ON REINFORCEMENT LEARNING
    Bi, Song
    Liu, Lei
    Han, Cunwu
    Sun, Dehui
    [J]. 2014 IEEE 3RD INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2014, : 347 - 350
  • [8] Improving reinforcement learning by using sequence trees
    Sertan Girgin
    Faruk Polat
    Reda Alhajj
    [J]. Machine Learning, 2010, 81 : 283 - 331
  • [9] Improving reinforcement learning by using sequence trees
    Girgin, Sertan
    Polat, Faruk
    Alhajj, Reda
    [J]. MACHINE LEARNING, 2010, 81 (03) : 283 - 331
  • [10] Composite rules selection using reinforcement learning for dynamic job-shop scheduling
    Wei, YZ
    Zhao, MY
    [J]. 2004 IEEE CONFERENCE ON ROBOTICS, AUTOMATION AND MECHATRONICS, VOLS 1 AND 2, 2004, : 1083 - 1088