Effective grasp detection method based on Swin transformer

被引:0
|
作者
Zhang, Jing [1 ,2 ]
Tang, Yulin [2 ]
Luo, Yusong [2 ]
Du, Yukun [2 ]
Chen, Mingju [3 ]
机构
[1] Univ Sci & Technol China, Dept Automat, Hefei, Peoples R China
[2] Southwest Univ Sci & Technol, Sch Informat Engn, Mianyang, Sichuan, Peoples R China
[3] Artificial Intelligence Key Lab Sichuan Prov, Yibin, Peoples R China
关键词
grasp detection; Swin transformer; attention mechanism; decoupled head; grasping tasks;
D O I
10.1117/1.JEI.33.3.033008
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
. Grasp detection within unstructured environments encounters challenges that lead to a reduced success rate in grasping attempts, attributable to factors including object uncertainty, random positions, and differences in perspective. This work proposes a grasp detection algorithm framework, Swin-transNet, which adopts a hypothesis treating graspable objects as a generalized category and distinguishing between graspable and non-graspable objects. The utilization of the Swin transformer module in this framework augments the feature extraction process, enabling the capture of global relationships within images. Subsequently, the integration of a decoupled head with attention mechanisms further refines the channel and spatial representation of features. This strategic combination markedly improves the system's adaptability to uncertain object categories and random positions, culminating in the precise output of grasping information. Moreover, we elucidate their roles in grasping tasks. We evaluate the grasp detection framework using the Cornell grasp dataset, which is divided into image and object levels. The experiment indicated a detection accuracy of 98.1% and a detection speed of 52 ms. Swin-transNet shows robust generalization on the Jacquard dataset, attaining a detection accuracy of 95.2%. It demonstrates an 87.8% success rate in real-world grasping testing on a visual grasping system, confirming its effectiveness for robotic grasping tasks.
引用
收藏
页数:22
相关论文
共 50 条
  • [31] Speech Keyword Spotting Method Based on Swin-Transformer Model
    Sun, Chengli
    Chen, Bikang
    Chen, Feilong
    Leng, Yan
    Guo, Qiaosheng
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2024, 17 (01)
  • [32] Fault diagnosis method based on Swin Transformer with path aggregation networks
    Liu, Chenyu
    Li, Zhinong
    Xiong, Pengwei
    Gu, Fengshou
    Zhendong yu Chongji/Journal of Vibration and Shock, 2024, 43 (18): : 258 - 266
  • [33] An efficient swin transformer-based method for underwater image enhancement
    Wang, Rong
    Zhang, Yonghui
    Zhang, Jian
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (12) : 18691 - 18708
  • [34] Swin Transformer for Pedestrian and Occluded Pedestrian Detection
    Liang, Jung-An
    Ding, Jian-Jiun
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [35] Remote Sensing Image Fusion Method Based on Improved Swin Transformer
    Li Zitong
    Zhao Jiankang
    Xu Jingran
    Long Haihui
    Liu Chuanqi
    ACTA PHOTONICA SINICA, 2023, 52 (11)
  • [36] Deep Supervised Hashing Image Retrieval Method Based on Swin Transformer
    Miao Z.
    Zhao X.
    Li Y.
    Wang J.
    Zhang R.
    Hunan Daxue Xuebao/Journal of Hunan University Natural Sciences, 2023, 50 (08): : 62 - 71
  • [37] A Swin transformer and MLP based method for identifying cherry ripeness and decay
    Song, Ke
    Yang, Jiwen
    Wang, Guohui
    FRONTIERS IN PHYSICS, 2023, 11
  • [38] Speech Keyword Spotting Method Based on Swin-Transformer Model
    Chengli Sun
    Bikang Chen
    Feilong Chen
    Yan Leng
    Qiaosheng Guo
    International Journal of Computational Intelligence Systems, 17
  • [39] Colorectal polyp segmentation method based on the Swin Transformer and graph reasoning
    Liang L.
    He A.
    Yang Y.
    Wu J.
    Gongcheng Kexue Xuebao/Chinese Journal of Engineering, 2024, 46 (05): : 897 - 907
  • [40] An efficient swin transformer-based method for underwater image enhancement
    Rong Wang
    Yonghui Zhang
    Jian Zhang
    Multimedia Tools and Applications, 2023, 82 : 18691 - 18708