Vision Transformer Based Multi-class Lesion Detection in IVOCT

被引:0
|
作者
Wang, Zixuan [1 ]
Shao, Yifan [2 ]
Sun, Jingyi [2 ]
Huang, Zhili [1 ]
Wang, Su [1 ]
Li, Qiyong [3 ]
Li, Jinsong [3 ]
Yu, Qian [2 ]
机构
[1] Sichuan Univ, Chengdu, Peoples R China
[2] Beihang Univ, Beijing, Peoples R China
[3] Sichuan Prov Peoples Hosp, Chengdu, Peoples R China
关键词
IVOCT; Object Detection; Vision Transformer;
D O I
10.1007/978-3-031-43987-2_32
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cardiovascular disease is a high-fatality illness. Intravascular Optical Coherence Tomography (IVOCT) technology can significantly assist in diagnosing and treating cardiovascular diseases. However, locating and classifying lesions from hundreds of IVOCT images is time-consuming and challenging, especially for junior physicians. An automatic lesion detection and classification model is desirable. To achieve this goal, in this work, we first collect an IVOCT dataset, including 2,988 images from 69 IVOCT data and 4,734 annotations of lesions spanning over three categories. Based on the newly-collected dataset, we propose a multi-class detection model based on Vision Transformer, called G-Swin Transformer. The essential part of our model is grid attention which is used to model relations among consecutive IVOCT images. Through extensive experiments, we show that the proposed G-Swin Transformer can effectively localize different types of lesions in IVOCT images, significantly outperforming baseline methods in all evaluation metrics. Our code is available via this link. https://github.com/Shao1Fan/G-Swin-Transformer
引用
收藏
页码:327 / 336
页数:10
相关论文
共 50 条
  • [21] The Multi-class SVM is Applied in Transformer Fault Diagnosis
    Qu, Liping
    Zhou, Haohan
    14TH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED COMPUTING AND APPLICATIONS FOR BUSINESS, ENGINEERING AND SCIENCE (DCABES 2015), 2015, : 477 - 480
  • [22] Multi-Class Intrusion Detection Based on Transformer for IoT Networks Using CIC-IoT-2023 Dataset
    Tseng, Shu-Ming
    Wang, Yan-Qi
    Wang, Yung-Chung
    FUTURE INTERNET, 2024, 16 (08)
  • [23] An adaptive Bagging algorithm based on lightweight transformer for multi-class imbalance recognition
    Wang, Junyi
    Jiang, Xuezheng
    Liu, Hailian
    Cai, Haibin
    Meng, Qinggang
    MULTIMEDIA SYSTEMS, 2024, 30 (02)
  • [24] An adaptive Bagging algorithm based on lightweight transformer for multi-class imbalance recognition
    Junyi Wang
    Xuezheng Jiang
    Hailian Liu
    Haibin Cai
    Qinggang Meng
    Multimedia Systems, 2024, 30
  • [25] Transformer-based active learning for multi-class text annotation and classification
    Afzal, Muhammad
    Hussain, Jamil
    Abbas, Asim
    Hussain, Maqbool
    Attique, Muhammad
    Lee, Sungyoung
    DIGITAL HEALTH, 2024, 10
  • [26] CCDN-DETR: A Detection Transformer Based on Constrained Contrast Denoising for Multi-Class Synthetic Aperture Radar Object Detection
    Zhang, Lei
    Zheng, Jiachun
    Li, Chaopeng
    Xu, Zhiping
    Yang, Jiawen
    Wei, Qiuxin
    Wu, Xinyi
    SENSORS, 2024, 24 (06)
  • [27] A fine-tuned vision transformer based enhanced multi-class brain tumor classification using MRI scan imagery
    Reddy, C. Kishor Kumar
    Reddy, Pulakurthi Anaghaa
    Janapati, Himaja
    Assiri, Basem
    Shuaib, Mohammed
    Alam, Shadab
    Sheneamer, Abdullah
    FRONTIERS IN ONCOLOGY, 2024, 14
  • [28] Scalable Multi-class Object Detection
    Razavi, Nima
    Gall, Juergen
    Van Gool, Luc
    2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011, : 1505 - 1512
  • [29] MULTI-CLASS FRUIT DETECTION BASED ON MULTIPLE COLOR CHANNELS
    Kuang, Hu-Lin
    Chan, Leanne Lai Hang
    Yan, Hong
    PROCEEDINGS OF 2015 INTERNATIONAL CONFERENCE ON WAVELET ANALYSIS AND PATTERN RECOGNITION (ICWAPR), 2015, : 1 - 7
  • [30] A Diffusion-Based Framework for Multi-Class Anomaly Detection
    He, Haoyang
    Zhang, Jiangning
    Chen, Hongxu
    Chen, Xuhai
    Li, Zhishan
    Chen, Xu
    Wang, Yabiao
    Wang, Chengjie
    Xie, Lei
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 8, 2024, : 8472 - 8480