ConvAttenMixer: Brain tumor detection and type classification using convolutional mixer with external and self-attention mechanisms

被引:12
|
作者
Alzahrani, Salha M. [1 ]
机构
[1] Taif Univ, Coll Comp & Informat Technol, Dept Comp Sci, POB 11099, Taif 21944, Saudi Arabia
关键词
Brain tumor classification; Transformer; Self-attention; External attention; Convolutional mixer; SEGMENTATION;
D O I
10.1016/j.jksuci.2023.101810
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Attention-based methods have recently demonstrated notable advancements in brain tumor classification. To further advance and strengthen this development, we have developed ConvAttenMixer, a transformer model that incorporates convolutional layers along with two attention mechanisms: self-attention and external attention. The proposed model utilizes two blocks of convolution mixers to effectively process and blend across patches, thereby enhancing the model's ability to capture spatial and channel-wise dependencies in MRI brain images. The self-attention block enables the model to prioritize important regions within the image and establish dependencies by assigning weights to each part based on their relevance to the task. This allows the model to emphasize crucial local features, disregard irrelevant ones, and capture interactions between different patches. On the other hand, the external attention block focuses more on significant global features and captures interactions among different images, enabling the model to establish dependencies and correlations across all samples. The classification head in the proposed model is a simple yet effective block designed to process the output feature maps using a squeeze-and-excitation mechanism, which in turn assigns higher weights to important channels and suppresses less-relevant channels. For experimentation, our ConvAttenMixer model was trained on a dataset consisting of 5712 MRI scans and subsequently tested on 1311 scans for classification into glioma, meningioma, pituitary tumor, and no-tumor images. Different variants of the proposed model were tested and evaluated. The optimally performing architecture was evaluated against the state-of-the-art baselines, namely self-attention MLP, external attention MLP, attention-based pooling convolutional net, and convolutional mixer net. Extensive experiments demonstrated that ConvAttenMixer outperformed the other baselines, which employed either self-attention or external attention mechanisms, while requiring significantly less computational memory. The suggested model exhibited higher precision, recall, and f-measure, achieving the highest accuracy of 0.9794 compared with the baselines' accuracy, which ranged from 0.87 to 0.93. The ConvAttenMixer model demonstrates the ability to operate locally on the patch level using self-attention and globally on the sample level using external attention, as well as prioritize important information on the spatial level and channel level using convolution mixers and the squeeze-and-excitation mechanism.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Image Classification based on Self-attention Convolutional Neural Network
    Cai, Xiaohong
    Li, Ming
    Cao, Hui
    Ma, Jingang
    Wang, Xiaoyan
    Zhuang, Xuqiang
    SIXTH INTERNATIONAL WORKSHOP ON PATTERN RECOGNITION, 2021, 11913
  • [2] Assessing the Impact of Attention and Self-Attention Mechanisms on the Classification of Skin Lesions
    Pedro, Rafael
    Oliveira, Arlindo L.
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [3] Bearing Fault Detection Based on Convolutional Self-Attention Mechanism
    Ye, Ruida
    Wang, Weijie
    Ren, Yuan
    Zhang, Keming
    PROCEEDINGS OF 2020 IEEE 2ND INTERNATIONAL CONFERENCE ON CIVIL AVIATION SAFETY AND INFORMATION TECHNOLOGY (ICCASIT), 2020, : 869 - 873
  • [4] Biscale Convolutional Self-Attention Network for Hyperspectral Coastal Wetlands Classification
    Luo, Junshen
    He, Zhi
    Lin, Haomei
    Wu, Heqian
    IEEE Geoscience and Remote Sensing Letters, 2024, 21 : 1 - 5
  • [5] Biscale Convolutional Self-Attention Network for Hyperspectral Coastal Wetlands Classification
    Luo, Junshen
    He, Zhi
    Lin, Haomei
    Wu, Heqian
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [6] Fine-grained entity type classification using GRU with self-attention
    Dhrisya K.
    Remya G.
    Mohan A.
    International Journal of Information Technology, 2020, 12 (3) : 869 - 878
  • [7] Fake news detection and classification using hybrid BiLSTM and self-attention model
    Mohapatra, Asutosh
    Thota, Nithin
    Prakasam, P.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (13) : 18503 - 18519
  • [8] Fake news detection and classification using hybrid BiLSTM and self-attention model
    Asutosh Mohapatra
    Nithin Thota
    P. Prakasam
    Multimedia Tools and Applications, 2022, 81 : 18503 - 18519
  • [9] Time Series Anomaly Detection in Vehicle Sensors Using Self-Attention Mechanisms
    Zhang, Ze
    Yao, Yue
    Hutabarat, Windo
    Farnsworth, Michael
    Tiwari, Divya
    Tiwari, Ashutosh
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (11) : 15964 - 15976
  • [10] Self-Attention Mechanisms-Based Laryngoscopy Image Classification Technique for Laryngeal Cancer Detection
    Kang, Yi-Fan
    Yang, Lie
    Hu, Yi-Fan
    Xu, Kai
    Cai, Lan-Jun
    Hu, Bin-Bin
    Lu, Xiang
    HEAD AND NECK-JOURNAL FOR THE SCIENCES AND SPECIALTIES OF THE HEAD AND NECK, 2024,