Learning Sliding Policy of Flat Multi-target Objects in Clutter Scenes

被引:0
|
作者
Wu, Liangdong [1 ]
Wu, Jiaxi [2 ]
Li, Zhengwei [3 ]
Chen, Yurou [2 ]
Liu, Zhiyong [1 ,2 ,4 ]
机构
[1] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
[3] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[4] Chinese Acad Sci, Cloud Comp Ctr, Dongguan, Guangdong, Peoples R China
来源
INFORMATION TECHNOLOGY AND CONTROL | 2024年 / 53卷 / 01期
关键词
Deep Learning in Manipulation; Reinforcement Learning; Robot Control; Intelligent system; sliding policy;
D O I
10.5755/j01.itc.53.1.34708
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In clutter scenes, one or several targets need to be obtained, which is hard for robot manipulation task. Especially, when the targets are flat objects like book, plates, due to limitation of common robot end-effectors, it will be more challenging. By employing pre-grasp operation like sliding, it becomes feasible to rearrange objects and shift the target towards table edge, enabling the robot to grasp it from a lateral perspective. In this paper, the proposed method transfers the task into a Parameterized Action Markov Decision Process to solve the problem, which is based on deep reinforcement learning. The mask images are taken as one of observations to the network for avoiding the impact of noise of original image. In order to improve data utilization, the policy network predicts the parameters for the sliding primitive of each object, which is weight-sharing, and then the Q-network selects the optimal execution target. Meanwhile, extra reward mechanism is adopted for improving the efficiency of task actions to cope with multiple targets. In addition, an adaptive policy scaling algorithm is proposed to improve the speed and adaptability of policy training. In both simulation and real system, our method achieves a higher task success rate and requires fewer actions to accomplish the flat multi-target sliding manipulation task within clutter scene, which verifies the effectiveness of ours.
引用
收藏
页码:5 / 18
页数:14
相关论文
共 50 条
  • [1] Multi-target tracking in clutter
    Sanders-Reed, JN
    Duncan, MJ
    Boucher, WB
    Dimmler, WM
    O'Keefe, S
    LASER WEAPONS TECHNOLOGY III, 2002, 4724 : 30 - 36
  • [2] Learning Pre-Grasp Manipulation of Multiple Flat Target Objects in Clutter
    Wu, Liangdong
    Wu, Jiaxi
    Chen, Yurou
    Li, Zhengwei
    Liu, Zhiyong
    2023 9TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS, ICCAR, 2023, : 371 - 376
  • [3] Multi-target multi-scan smoothing in clutter
    Kim, Tae Han
    Song, Taek Lyul
    IET RADAR SONAR AND NAVIGATION, 2016, 10 (07): : 1270 - 1276
  • [4] Modelling multi-target estimation in noise and clutter
    Malyutov, MB
    Tsitovich, II
    SIMULATION IN INDUSTRY'2000, 2000, : 598 - 600
  • [5] Multi-Target Track Initiation in Heavy Clutter
    Xu, Li
    Lou, Ruzhen
    Zhang, Chuanbin
    Lang, Bo
    Ding, Weiyue
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 72 (03): : 4489 - 4507
  • [6] Multi-target tracking in clutter without measurement assignment
    Musicki, D
    La Scala, BR
    Evans, RJ
    2004 43RD IEEE CONFERENCE ON DECISION AND CONTROL (CDC), VOLS 1-5, 2004, : 716 - 721
  • [7] Region clutter estimation method for multi-target tracking
    Liu, G. (gxliu@xidian.edu.cn), 1600, Chinese Society of Astronautics (35):
  • [8] Multi-Target Tracking in Clutter without Measurement Assignment
    Musicki, Darko
    La Scala, Barbara
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2008, 44 (03) : 877 - 896
  • [9] Multi-Target Tracking in Clutter Aided by Doppler Information
    Jin B.
    Li C.
    Guo J.
    He D.-J.
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2019, 48 (04): : 511 - 517
  • [10] Multi-target tracking in clutter with sequential Monte Carlo methods
    Liu, B.
    Ji, C.
    Zhang, Y.
    Hao, C.
    Wong, K. -K.
    IET RADAR SONAR AND NAVIGATION, 2010, 4 (05): : 662 - 672