Attention-based efficient robot grasp detection network

被引:3
|
作者
Qin, Xiaofei [1 ]
Hu, Wenkai [1 ]
Xiao, Chen [2 ]
He, Changxiang [2 ]
Pei, Songwen [1 ,3 ,4 ]
Zhang, Xuedian [1 ,3 ,4 ,5 ]
机构
[1] Univ Shanghai Sci & Technol, Sch Opt Elect & Comp Engn, Shanghai 200093, Peoples R China
[2] Univ Shanghai Sci & Technol, Coll Sci, Shanghai 200093, Peoples R China
[3] Shanghai Key Lab Modern Opt Syst, Shanghai 200093, Peoples R China
[4] Minist Educ, Key Lab Biomed Opt Technol & Devices, Shanghai 200093, Peoples R China
[5] Tongji Univ, Shanghai Inst Intelligent Sci & Technol, Shanghai 201210, Peoples R China
关键词
Robot grasp detection; Attention mechanism; Encoder-decoder; Neural network;
D O I
10.1631/FITEE.2200502
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To balance the inference speed and detection accuracy of a grasp detection algorithm, which are both important for robot grasping tasks, we propose an encoder-decoder structured pixel-level grasp detection neural network named the attention-based efficient robot grasp detection network (AE-GDN). Three spatial attention modules are introduced in the encoder stages to enhance the detailed information, and three channel attention modules are introduced in the decoder stages to extract more semantic information. Several lightweight and efficient DenseBlocks are used to connect the encoder and decoder paths to improve the feature modeling capability of AE-GDN. A high intersection over union (IoU) value between the predicted grasp rectangle and the ground truth does not necessarily mean a high-quality grasp configuration, but might cause a collision. This is because traditional IoU loss calculation methods treat the center part of the predicted rectangle as having the same importance as the area around the grippers. We design a new IoU loss calculation method based on an hourglass box matching mechanism, which will create good correspondence between high IoUs and high-quality grasp configurations. AEGDN achieves the accuracy of 98.9% and 96.6% on the Cornell and Jacquard datasets, respectively. The inference speed reaches 43.5 frames per second with only about 1.2 x 10(6) parameters. The proposed AE-GDN has also been deployed on a practical robotic arm grasping system and performs grasping well.
引用
收藏
页码:1430 / 1444
页数:15
相关论文
共 50 条
  • [21] Reverse Attention-Based Residual Network for Salient Object Detection
    Chen, Shuhan
    Tan, Xiuli
    Wang, Ben
    Lu, Huchuan
    Hu, Xuelong
    Fu, Yun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3763 - 3776
  • [22] Attention-based Convolutional Neural Network for ASV Spoofing Detection
    Ling, Hefei
    Huang, Leichao
    Huang, Junrui
    Zhang, Baiyan
    Li, Ping
    INTERSPEECH 2021, 2021, : 4289 - 4293
  • [23] Towards an efficient contextual perception for humanoid robot: A selective attention-based approach
    Jiang, Yanrong
    Xiao, Nanfeng
    Zhang, Linying
    WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 65 - 65
  • [24] A Robot Grasp Relationship Detection Network Based on the Fusion of Multiple Features
    Chi, Jianning
    Wu, Xingrui
    Ma, Changqing
    Yu, Xiaosheng
    Wu, Chengdong
    PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 1479 - 1484
  • [25] A pixel-level grasp detection method based on Efficient Grasp Aware Network
    Xi, Haonan
    Li, Shaodong
    Liu, Xi
    ROBOTICA, 2024, 42 (09) : 3190 - 3210
  • [26] Efficient Grasp Detection Network With Gaussian-Based Grasp Representation for Robotic Manipulation
    Cao, Hu
    Chen, Guang
    Li, Zhijun
    Feng, Qian
    Lin, Jianjie
    Knoll, Alois
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2023, 28 (03) : 1384 - 1394
  • [27] ADCGNet: Attention-based dual channel Gabor network towards efficient detection and classification of electrocardiogram images
    Arhin, Joseph Roger
    Zhang, Xiaoling
    Coker, Kenneth
    Agyemang, Isaac Osei
    Attipoe, Wisdom Kwame
    Sam, Francis
    Adjei-Mensah, Isaac
    Agyei, Emmanuel
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (09)
  • [28] An Efficient Attention-Based Network for Screening Major Depressive Disorder with sMRI
    Qu, Xiaohan
    Xiong, Yuyang
    Zhai, Kai
    Yang, Xiaoyu
    Yang, Jun
    2023 29TH INTERNATIONAL CONFERENCE ON MECHATRONICS AND MACHINE VISION IN PRACTICE, M2VIP 2023, 2023,
  • [29] Attention-Based Multiscale Feature Fusion for Efficient Surface Defect Detection
    Zhao, Yuhao
    Liu, Qing
    Su, Hu
    Zhang, Jiabin
    Ma, Hongxuan
    Zou, Wei
    Liu, Song
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 10
  • [30] Attention-based Multiscale Context Awareness Network for Insulator Defect Detection
    Zeng, Junting
    Zhu, Xinshan
    Li, Bin
    Guo, Zhimin
    Tian, Yangyang
    Yuan, Shaoguang
    2022 4TH INTERNATIONAL CONFERENCE ON SMART POWER & INTERNET ENERGY SYSTEMS, SPIES, 2022, : 2350 - 2355