HTC-Grasp: A Hybrid Transformer-CNN Architecture for Robotic Grasp Detection

被引:4
|
作者
Zhang, Qiang [1 ]
Zhu, Jianwei [1 ]
Sun, Xueying [1 ]
Liu, Mingmin [2 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Automat, 666 Changhui Rd, Zhenjiang 212100, Peoples R China
[2] SIASUN Robot & Automat Co Ltd, Cent Res Inst, 16 Jinhui St, Shenyang 110168, Peoples R China
基金
中国国家自然科学基金;
关键词
robotic grasp; transformer; attentional mechanism;
D O I
10.3390/electronics12061505
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Accurately detecting suitable grasp areas for unknown objects through visual information remains a challenging task. Drawing inspiration from the success of the Vision Transformer in vision detection, the hybrid Transformer-CNN architecture for robotic grasp detection, known as HTC-Grasp, is developed to improve the accuracy of grasping unknown objects. The architecture employs an external attention-based hierarchical Transformer as an encoder to effectively capture global context and correlation features across the entire dataset. Furthermore, a channel-wise attention-based CNN decoder is presented to adaptively adjust the weight of the channels in the approach, resulting in more efficient feature aggregation. The proposed method is validated on the Cornell and the Jacquard dataset, achieving an image-wise detection accuracy of 98.3% and 95.8% on each dataset, respectively. Additionally, the object-wise detection accuracy of 96.9% and 92.4% on the same datasets are achieved based on this method. A physical experiment is also performed using the Elite 6Dof robot, with a grasping accuracy rate of 93.3%, demonstrating the proposed method's ability to grasp unknown objects in real scenarios. The results of this study indicate that the proposed method outperforms other state-of-the-art methods.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] SKGNet: Robotic Grasp Detection With Selective Kernel Convolution
    Yu, Sheng
    Zhai, Di-Hua
    Xia, Yuanqing
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2023, 20 (04) : 2241 - 2252
  • [32] Jacquard: A Large Scale Dataset for Robotic Grasp Detection
    Depierre, Amaury
    Dellandrea, Emmanuel
    Chen, Liming
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 3511 - 3516
  • [33] Robotic Grasp Pose Detection Using Deep Learning
    Caldera, Shehan
    Rassau, Alexander
    Chai, Douglas
    2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 1966 - 1972
  • [34] Robotic Grasp Detection Using Extreme Learning Machine
    Sun, Changliang
    Yu, Yuanlong
    Liu, Huaping
    Gu, Jason
    2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2015, : 1115 - 1120
  • [35] Single-Grasp Detection Based on Rotational Region CNN
    Jiang, Shan
    Zhao, Xi
    Cai, Zhenhua
    Xiang, Kui
    Ju, Zhaojie
    ADVANCES IN COMPUTATIONAL INTELLIGENCE SYSTEMS (UKCI 2019), 2020, 1043 : 131 - 141
  • [36] Transformer-CNN hybrid network for improving PET time of flight prediction
    Feng, Xuhui
    Muhashi, Amanjule
    Onishi, Yuya
    Ota, Ryosuke
    Liu, Huafeng
    PHYSICS IN MEDICINE AND BIOLOGY, 2024, 69 (11):
  • [37] HGG-CNN: The Generation of the Optimal Robotic Grasp Pose Based on Vision
    Qiu, Shiyin
    Lodder, David
    Du, Feifan
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2020, 26 (06): : 1517 - 1529
  • [38] Hybrid transformer-CNN model for accurate prediction of peptide hemolytic potential
    Almotairi, Sultan
    Badr, Elsayed
    Abdelbaky, Ibrahim
    Elhakeem, Mohamed
    Abdul Salam, Mustafa
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [39] MTGrasp: Multiscale 6-DoF Robotic Grasp Detection
    Yu, Sheng
    Zhai, Di-Hua
    Xia, Yuanqing
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2024, : 1 - 12
  • [40] On the Importance of Label Encoding and Uncertainty Estimation for Robotic Grasp Detection
    Stephan, Benedict
    Aganian, Dustin
    Hinneburg, Lars
    Eisenbach, Markus
    Mueller, Steffen
    Gross, Horst-Michael
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 4781 - 4788