Unsupervised Visual Anomaly Detection Using Self-Supervised Pre-Trained Transformer

被引:0
|
作者
Kim, Jun-Hyung [1 ]
Kwon, Goo-Rak [1 ]
机构
[1] Chosun Univ, Dept Informat & Commun Engn, Gwangju 61452, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Image reconstruction; Image segmentation; Transformers; Computational modeling; Location awareness; Feature extraction; Anomaly detection; Data augmentation; Self-supervised learning; data-augmentation; self-supervised learning; transformer;
D O I
10.1109/ACCESS.2024.3454753
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the various industrial manufacturing processes, the automatic visual inspection system is an essential part as it reduces the chances of delivering defective products and the cost of training and hiring experts for manual inspection. In this work, we propose a new unsupervised anomaly detection method inspired by the masked language model for the automatic visual inspection system. The proposed method consists of an image tokenizer and two subnetworks, a reconstruction subnetwork, and a segmentation subnetwork. We adopt a pre-trained self-supervised vision Transformer model to use it as an image tokenizer. Our first subnetwork is trained to predict the anomaly-free patch tokens and the second subnetwork is trained to produce anomaly segmentation results from both the reconstructed and input patch tokens. During training, only the two subnetworks are optimized, and parameters of an image tokenizer are frozen. Experimental results show that the proposed method exhibits better performance than conventional methods in detecting defective products by achieving 99.05% I-AUROC on MVTecAD dataset and 94.8% I-AUROC on BTAD.
引用
收藏
页码:127604 / 127613
页数:10
相关论文
共 50 条
  • [2] Prediction of Protein Tertiary Structure Using Pre-Trained Self-Supervised Learning Based on Transformer
    Kurniawan, Alif
    Jatmiko, Wisnu
    Hertadi, Rukman
    Habibie, Novian
    [J]. 2020 5TH INTERNATIONAL WORKSHOP ON BIG DATA AND INFORMATION SECURITY (IWBIS 2020), 2020, : 75 - 80
  • [3] END-TO-END SPOKEN LANGUAGE UNDERSTANDING USING TRANSFORMER NETWORKS AND SELF-SUPERVISED PRE-TRAINED FEATURES
    Morais, Edmilson
    Kuo, Hong-Kwang J.
    Thomas, Samuel
    Tuske, Zoltan
    Kingsbury, Brian
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7483 - 7487
  • [4] Speech Enhancement Using Self-Supervised Pre-Trained Model and Vector Quantization
    Zhao, Xiao-Ying
    Zhu, Qiu-Shi
    Zhang, Jie
    [J]. PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 330 - 334
  • [5] Self-Supervised Masking for Unsupervised Anomaly Detection and Localization
    Huang, Chaoqin
    Xu, Qinwei
    Wang, Yanfeng
    Wang, Yu
    Zhang, Ya
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4426 - 4438
  • [6] Explore the Use of Self-supervised Pre-trained Acoustic Features on Disguised Speech Detection
    Quan, Jie
    Yang, Yingchun
    [J]. BIOMETRIC RECOGNITION (CCBR 2021), 2021, 12878 : 483 - 490
  • [7] A Systematic Review of Transformer-Based Pre-Trained Language Models through Self-Supervised Learning
    Kotei, Evans
    Thirunavukarasu, Ramkumar
    [J]. INFORMATION, 2023, 14 (03)
  • [8] SPIQ: A Self-Supervised Pre-Trained Model for Image Quality Assessment
    Chen, Pengfei
    Li, Leida
    Wu, Qingbo
    Wu, Jinjian
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 513 - 517
  • [9] BadEncoder: Backdoor Attacks to Pre-trained Encoders in Self-Supervised Learning
    Jia, Jinyuan
    Liu, Yupei
    Gong, Neil Zhenqiang
    [J]. 43RD IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2022), 2022, : 2043 - 2059
  • [10] Self-Supervised Quantization of Pre-Trained Neural Networks for Multiplierless Acceleration
    Vogel, Sebastian
    Springer, Jannik
    Guntoro, Andre
    Ascheid, Gerd
    [J]. 2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 1094 - 1099