Enhance fashion classification of mosquito vector species via self-supervised vision transformer

被引:0
|
作者
Veerayuth Kittichai [1 ]
Morakot Kaewthamasorn [2 ]
Tanawat Chaiphongpachara [3 ]
Sedthapong Laojun [3 ]
Tawee Saiwichai [4 ]
Kaung Myat Naing [6 ]
Teerawat Tongloy [6 ]
Siridech Boonsang [5 ]
Santhad Chuwongin [6 ]
机构
[1] King Mongkut’s Institute of Technology Ladkrabang,Faculty of Medicine
[2] Chulalongkorn University,Veterinary Parasitology Research Unit, Faculty of Veterinary Science
[3] Suan Sunandha Rajabhat University,Department of Public Health and Health Promotion, College of Allied Health Science
[4] Mahidol University,Department of Parasitology and Entomology, Faculty of Public Health
[5] King Mongkut’s Institute of Technology Ladkrabang,Department of Electrical Engineering, School of Engineering
[6] King Mongkut’s Institute of Technology Ladkrabang,College of Advanced Manufacturing Innovation
关键词
Mosquito vector species; Artificial intelligence; Self-distillation with unlabeled data; Mobile phone application;
D O I
10.1038/s41598-024-83358-8
中图分类号
学科分类号
摘要
Vector-borne diseases pose a major worldwide health concern, impacting more than 1 billion people globally. Among various blood-feeding arthropods, mosquitoes stand out as the primary carriers of diseases significant in both medical and veterinary fields. Hence, comprehending their distinct role fulfilled by different mosquito types is crucial for efficiently addressing and enhancing control measures against mosquito-transmitted diseases. The conventional method for identifying mosquito species is laborious and requires significant effort to learn. Classification is subsequently carried out by skilled laboratory personnel, rendering the process inherently time-intensive and restricting the task to entomology specialists. Therefore, integrating artificial intelligence with standard taxonomy, such as molecular techniques, is essential for accurate mosquito species identification. Advancement in novel tools with artificial intelligence has challenged the task of developing an automated system for sample collection and identification. This study aims to introduce a self-supervised Vision Transformer supporting an automatic model for classifying mosquitoes found across various regions of Thailand. The objective is to utilize self-distillation with unlabeled data (DINOv2) to develop models on a mobile phone-captured dataset containing 16 species of female mosquitoes, including those known for transmitting malaria and dengue. The DINOv2 model surpassed the ViT baseline model in precision and recall for all mosquito species. When compared on a species-specific level, utilizing the DINOv2 model resulted in reductions in false negatives and false positives, along with enhancements in precision and recall values, in contrast to the baseline model, across all mosquito species. Notably, at least 10 classes exhibited outstanding performance, achieving above precision and recall rates exceeding 90%. Remarkably, when applying cropping techniques to the dataset instead of utilizing the original photographs, there was a significant improvement in performance across all DINOv2 models studied. This is demonstrated by an increase in recall to 87.86%, precision to 91.71%, F1 score to 88.71%, and accuracy to 98.45%, respectively. Malaria mosquito species can be easily distinguished from another genus like Aedes, Mansonia, Armigeres, and Culex, respectively. While classifying malaria vector species presented challenges for the DINOv2 model, utilizing the cropped images enhanced precision by up to 96% for identifying one of the top three malaria vectors in Thailand, Anopheles minimus. A proficiently trained DINOv2 model, coupled with effective data management, can contribute to the development of a mobile phone application. Furthermore, this method shows promise in supporting field professionals who are not entomology experts in effectively addressing pathogens responsible for diseases transmitted by female mosquitoes.
引用
下载
收藏
相关论文
共 50 条
  • [41] EVALUATING CONVNET AND TRANSFORMER BASED SELF-SUPERVISED ALGORITHMS FOR BUILDING ROOF FORM CLASSIFICATION
    Mutreja, G.
    Bittner, K.
    GEOSPATIAL WEEK 2023, VOL. 48-1, 2023, : 315 - 321
  • [42] Self-supervised vision transformers for semantic segmentation
    Gu, Xianfan
    Hu, Yingdong
    Wen, Chuan
    Gao, Yang
    Computer Vision and Image Understanding, 2025, 251
  • [43] Distilling Self-Supervised Vision Transformers for Weakly-Supervised Few-Shot Classification & Segmentation
    Kang, Dahyun
    Koniusz, Piotr
    Cho, Minsu
    Murray, Naila
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 19627 - 19638
  • [44] Emerging Properties in Self-Supervised Vision Transformers
    Caron, Mathilde
    Touvron, Hugo
    Misra, Ishan
    Jegou, Herve
    Mairal, Julien
    Bojanowski, Piotr
    Joulin, Armand
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9630 - 9640
  • [45] Clinical Outcome Prediction in COVID-19 using Self-supervised Vision Transformer Representations
    Konwer, Aishik
    Prasanna, Prateek
    MEDICAL IMAGING 2022: COMPUTER-AIDED DIAGNOSIS, 2022, 12033
  • [46] Self-Supervised Vision Transformers for Malware Detection
    Seneviratne, Sachith
    Shariffdeen, Ridwan
    Rasnayaka, Sanka
    Kasthuriarachchi, Nuran
    IEEE ACCESS, 2022, 10 : 103121 - 103135
  • [47] Scaling Vision Transformers to Gigapixel Images via Hierarchical Self-Supervised Learning
    Chen, Richard J.
    Chen, Chengkuan
    Li, Yicong
    Chen, Tiffany Y.
    Trister, Andrew D.
    Krishnan, Rahul G.
    Mahmood, Faisal
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16123 - 16134
  • [48] Personvit: large-scale self-supervised vision transformer for person re-identification
    Hu, Bin
    Wang, Xinggang
    Liu, Wenyu
    Machine Vision and Applications, 2025, 36 (02)
  • [49] Few-shot segmentation for esophageal OCT images based on self-supervised vision transformer
    Wang, Cong
    Gan, Meng
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2024, 34 (02)
  • [50] Self-Supervised RGB-NIR Fusion Video Vision Transformer Framework for rPPG Estimation
    Park, Soyeon
    Kim, Bo-Kyeong
    Dong, Suh-Yeon
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71