Spiking ViT: spiking neural networks with transformer-attention for steel surface defect classification

被引:2
|
作者
Gong, Liang [1 ]
Dong, Hang [1 ]
Zhang, Xinyu [1 ]
Cheng, Xin [2 ]
Ye, Fan [3 ]
Guo, Liangchao [1 ]
Ge, Zhenghui [1 ]
机构
[1] Yangzhou Univ, Sch Mech Engn, Yangzhou, Jiangsu, Peoples R China
[2] Yangzhou Univ, Sch Informat Engn, Sch Artificial Intelligence, Yangzhou, Jiangsu, Peoples R China
[3] Tongling Nonferrous Met Grp Co Ltd, Tongling, Peoples R China
基金
中国博士后科学基金;
关键词
surface defect classification; vision transformer; spiking neural network; deep learning; data modeling;
D O I
10.1117/1.JEI.33.3.033001
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
. Throughout the steel production process, a variety of surface defects inevitably occur. These defects can impair the quality of steel products and reduce manufacturing efficiency. Therefore, it is crucial to study and categorize the multiple defects on the surface of steel strips. Vision transformer (ViT) is a unique neural network model based on a self-attention mechanism that is widely used in many different disciplines. Conventional ViT ignores the specifics of brain signaling and instead uses activation functions to simulate genuine neurons. One of the fundamental building blocks of a spiking neural network is leaky integration and fire (LIF), which has biodynamic characteristics akin to those of a genuine neuron. LIF neurons work in an event-driven manner such that higher performance can be achieved with less power. The goal of this work is to integrate ViT and LIF neurons to build and train an end-to-end hybrid network architecture, spiking vision transformer (S-ViT), for the classification of steel surface defects. The framework relies on the ViT architecture by replacing the activation functions used in ViT with LIF neurons, constructing a global spike feature fusion module spiking transformer encoder as well as a spiking-MLP classification head for implementing the classification functionality and using it as a basic building block of S-ViT. Based on the experimental results, our method has demonstrated outstanding classification performance across all metrics. The overall test accuracies of S-ViT are 99.41%, 99.65%, 99.54%, and 99.77% on NEU-CLSs, and 95.70%, 95.93%, 96.94%, and 97.19% on XSDD. S-ViT achieves superior classification performance compared to convolutional neural networks and recent findings. Its performance is also improved relative to the original ViT model. Furthermore, the robustness test results of S-ViT show that S-ViT still maintains reliable accuracy when recognizing images that contain Gaussian noise.
引用
收藏
页数:20
相关论文
共 50 条
  • [21] Integrating Non-spiking Interneurons in Spiking Neural Networks
    Strohmer, Beck
    Stagsted, Rasmus Karnoe
    Manoonpong, Poramate
    Larsen, Leon Bonde
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [22] Covert attention with a spiking neural network
    Chevallier, Sylvain
    Tarroux, Philippe
    COMPUTER VISION SYSTEMS, PROCEEDINGS, 2008, 5008 : 56 - 65
  • [23] Sa-SNN: spiking attention neural network for image classification
    Dan, Yongping
    Wang, Zhida
    Li, Hengyi
    Wei, Jintong
    PEERJ, 2024, 10 : 1 - 23
  • [24] Moving Target Detection and Classification Using Spiking Neural Networks
    Cai, Rongtai
    Wu, Qingxiang
    Wang, Ping
    Sun, Honghai
    Wang, Zichen
    INTELLIGENT SCIENCE AND INTELLIGENT DATA ENGINEERING, ISCIDE 2011, 2012, 7202 : 210 - 217
  • [25] Selective Input Sparsity in Spiking Neural Networks for Pattern Classification
    Leigh, Alexander J.
    Heidarpur, Moslem
    Mirhassani, Mitra
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 799 - 803
  • [26] Simulation of spiking neural networks
    Bako, Laszlo
    Szekely, Iuliu
    David, Laszlo
    Brassai, Tihamer Sandor
    PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON OPTIMIZATION OF ELECTRICAL AND ELECTRONIC EQUIPMENT, VOL III: INDUSTRIAL AUTOMATION AND CONTROL, 2004, : 179 - 184
  • [27] Agreement in Spiking Neural Networks
    Kunev, Martin
    Kuznetsov, Petr
    Sheynikhovich, Denis
    JOURNAL OF COMPUTATIONAL BIOLOGY, 2022, 29 (04) : 358 - 369
  • [28] A Survey on Spiking Neural Networks
    Han, Chan Sik
    Lee, Keon Myung
    INTERNATIONAL JOURNAL OF FUZZY LOGIC AND INTELLIGENT SYSTEMS, 2021, 21 (04) : 317 - 337
  • [29] Applications of spiking neural networks
    Bohte, SM
    Kok, JN
    INFORMATION PROCESSING LETTERS, 2005, 95 (06) : 519 - 520
  • [30] Classification of Persian Handwritten Digits Using Spiking Neural Networks
    Kiani, Kourosh
    Korayem, Elmira Mohsenzadeh
    2015 2ND INTERNATIONAL CONFERENCE ON KNOWLEDGE-BASED ENGINEERING AND INNOVATION (KBEI), 2015, : 1113 - 1116