An efficient approach to detect and segment underwater images using Swin Transformer

被引:2
|
作者
Pavithra, S. [1 ]
Denny, J. Cicil Melbin [1 ]
机构
[1] Vellore Inst Technol, Sch Comp Sci & Engn, Chennai, Tamil Nadu, India
关键词
Artificial Intelligence; Attention mechanisms; ConvMixer; Swin Transformer; Underwater environments;
D O I
10.1016/j.rineng.2024.102460
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In managing the fish environment and underwater ecosystem, there 's been a big shift from old-generational schoolings, hands-on methods to high-technologies, automatic ways of doing things, because of some clever computer technology, using a type of Artificial Intelligence smartness known as Convolutional Neural Networks, or CNNs in short. These smart systems help deal with the tricky task of keeping tabs on where fish stay out in deeper underwater, particularly when it 's hard to see underwater and when the fish are all over the place. There are some pretty good tools and techniques already out there, like DPANet and a few others with names like PIFSNet, EFS -Net, and MFAS-Net, which have been helping a bunch of researches and developments for underwater ecosystem. But Deeper water environment still needs a newer and better way to get the tasks done. So, to answer this major question, this hybrid architecture rolls out a new combination that takes the best bits of transformer architecture known as Swin Transformer and another mechanism known as ConvMixer, making it in a form of SwinUNet Architecture. This Swin Transformer is well defined at picking up on the spatial information from underwater deeper images in our case, which is a key point to be highlighted in really getting the fish 's niche turf and enquiring their ecosystem. And when the underwater turns misty, the specialized implementation of ConvMixer modules incorporated with Swin Transformer increases its ability to sort out the fish from the rest of the underwater world. And this makes it to perform even more better compared with the existing state of arts model, even under Modern Underwater Machine Vision System conditions. This model even over takes to utilization of Few-Shot Learning model for training it in this lesser Underwater Images, with its capability of identifying the patterns and differentiations in both larger and smaller class objects in the given ground truth image. In this new approach, the paper combines the smart focusing methods with a way to blend different signals to better mIoU score, making it super useful for having a deeper understanding on fish-ecosystem and also their underwater ecosystem. While comparing it with the pre-existing state of arts model and the booming Image Segmentation models (YOLO v8, Pix 2 Pix GAN model, Auto Encoders, UNet), the hybrid approach outperformed the existing model results on this Semantic Segmentation for Underwater Imagery (SUIM) dataset, which is been disclosed in the further sections below.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] SwinFuse: A Residual Swin Transformer Fusion Network for Infrared and Visible Images
    Wang, Zhishe
    Chen, Yanlin
    Shao, Wenyu
    Li, Hui
    Zhang, Lei
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [42] Image super-resolution reconstruction using Swin Transformer with efficient channel attention networks
    Sun, Zhenxi
    Zhang, Jin
    Chen, Ziyi
    Hong, Lu
    Zhang, Rui
    Li, Weishi
    Xia, Haojie
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 136
  • [43] Efficient brain tumor segmentation using Swin transformer and enhanced local self-attention
    Fethi Ghazouani
    Pierre Vera
    Su Ruan
    International Journal of Computer Assisted Radiology and Surgery, 2024, 19 : 273 - 281
  • [44] Efficient brain tumor segmentation using Swin transformer and enhanced local self-attention
    Ghazouani, Fethi
    Vera, Pierre
    Ruan, Su
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2023, 19 (2) : 273 - 281
  • [45] Efficient Stereo Matching Using Swin Transformer and Multilevel Feature Consistency in Autonomous Mobile Systems
    Su, Xiaojie
    Liu, Shimin
    Li, Rui
    Bing, Zhenshan
    Knoll, Alois
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (05) : 7957 - 7965
  • [46] Leveraging swin transformer with ensemble of deep learning model for cervical cancer screening using colposcopy images
    Himabindu, D. Dakshayani
    Lydia, E. Laxmi
    Rajesh, M. V.
    Ahmed, Mohammed Altaf
    Ishak, Mohamad Khairi
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [47] Classification of Muscular Dystrophies from MR Images Improves Using the Swin Transformer Deep Learning Model
    Mastropietro, Alfonso
    Casali, Nicola
    Taccogna, Maria Giovanna
    D'Angelo, Maria Grazia
    Rizzo, Giovanna
    Peruzzo, Denis
    BIOENGINEERING-BASEL, 2024, 11 (06):
  • [48] Implementation of Chaotic Atom Search Optimized SWIN Transformer Architecture for Efficient Corpus Callosum Segmentation in Brain MRI Images
    Avasarala, Padmanabha Sarma
    Govindakumar, Saranya
    PRZEGLAD ELEKTROTECHNICZNY, 2023, 99 (12): : 185 - 189
  • [49] Intelligent Wood Inspection Approach Utilizing Enhanced Swin Transformer
    Ding, Zhigang
    Fu, Fucheng
    Zheng, Jishi
    Yang, Haiyan
    Zou, Fumin
    Linghua, Kong
    IEEE ACCESS, 2024, 12 : 16794 - 16804
  • [50] Chromosome Cluster Type Identification Using a Swin Transformer
    Joshi, Indu
    Mondal, Arnab Kumar
    Navab, Nassir
    APPLIED SCIENCES-BASEL, 2023, 13 (14):