An efficient approach to detect and segment underwater images using Swin Transformer

被引:2
|
作者
Pavithra, S. [1 ]
Denny, J. Cicil Melbin [1 ]
机构
[1] Vellore Inst Technol, Sch Comp Sci & Engn, Chennai, Tamil Nadu, India
关键词
Artificial Intelligence; Attention mechanisms; ConvMixer; Swin Transformer; Underwater environments;
D O I
10.1016/j.rineng.2024.102460
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In managing the fish environment and underwater ecosystem, there 's been a big shift from old-generational schoolings, hands-on methods to high-technologies, automatic ways of doing things, because of some clever computer technology, using a type of Artificial Intelligence smartness known as Convolutional Neural Networks, or CNNs in short. These smart systems help deal with the tricky task of keeping tabs on where fish stay out in deeper underwater, particularly when it 's hard to see underwater and when the fish are all over the place. There are some pretty good tools and techniques already out there, like DPANet and a few others with names like PIFSNet, EFS -Net, and MFAS-Net, which have been helping a bunch of researches and developments for underwater ecosystem. But Deeper water environment still needs a newer and better way to get the tasks done. So, to answer this major question, this hybrid architecture rolls out a new combination that takes the best bits of transformer architecture known as Swin Transformer and another mechanism known as ConvMixer, making it in a form of SwinUNet Architecture. This Swin Transformer is well defined at picking up on the spatial information from underwater deeper images in our case, which is a key point to be highlighted in really getting the fish 's niche turf and enquiring their ecosystem. And when the underwater turns misty, the specialized implementation of ConvMixer modules incorporated with Swin Transformer increases its ability to sort out the fish from the rest of the underwater world. And this makes it to perform even more better compared with the existing state of arts model, even under Modern Underwater Machine Vision System conditions. This model even over takes to utilization of Few-Shot Learning model for training it in this lesser Underwater Images, with its capability of identifying the patterns and differentiations in both larger and smaller class objects in the given ground truth image. In this new approach, the paper combines the smart focusing methods with a way to blend different signals to better mIoU score, making it super useful for having a deeper understanding on fish-ecosystem and also their underwater ecosystem. While comparing it with the pre-existing state of arts model and the booming Image Segmentation models (YOLO v8, Pix 2 Pix GAN model, Auto Encoders, UNet), the hybrid approach outperformed the existing model results on this Semantic Segmentation for Underwater Imagery (SUIM) dataset, which is been disclosed in the further sections below.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Utilizing Swin Transformer for the Classification of Ophthalmic Diseases in Optical Coherence Tomography (OCT) Images: A Novel Approach
    Mapanao, Jay Ryan
    Luis Lozano, Paulo
    2024 6TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND THE INTERNET, ICCCI 2024, 2024, : 94 - 100
  • [32] Waterbody Detection for the Reservoirs in South Korea Using Swin Transformer and Sentinel-1 Images
    Choi, Soyeon
    Youn, Youjeong
    Kang, Jonggu
    Kim, Seoyeon
    Jeong, Yemin
    Im, Yungyo
    Seo, Youngmin
    Kim, Wanyub
    Choi, Minha
    Lee, Yangwon
    KOREAN JOURNAL OF REMOTE SENSING, 2023, 39 (5-3) : 949 - 965
  • [33] Source Microphone Identification Using Swin Transformer
    Qamhan, Mustafa
    Alotaibi, Yousef A.
    Selouani, Sid-Ahmed
    APPLIED SCIENCES-BASEL, 2023, 13 (12):
  • [34] Building Extraction from Satellite Images Using Mask R-CNN and Swin Transformer
    Gibril, Mohamed Barakat A.
    Al-Ruzouq, Rami
    Bolcek, Jan
    Shanableh, Abdallah
    Jena, Ratiranjan
    2024 34TH INTERNATIONAL CONFERENCE RADIOELEKTRONIKA, RADIOELEKTRONIKA 2024, 2024,
  • [35] SwinIR: Image Restoration Using Swin Transformer
    Liang, Jingyun
    Cao, Jiezhang
    Sun, Guolei
    Zhang, Kai
    Van Gool, Luc
    Timofte, Radu
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 1833 - 1844
  • [36] SwinSAM: Fine-grained polyp segmentation in colonoscopy images via segment anything model integrated with a Swin Transformer decoder
    Feng, Zhoushan
    Zhang, Yuliang
    Chen, Yanhong
    Shi, Yiyu
    Liu, Yu
    Sun, Wen
    Du, Lili
    Chen, Dunjin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 100
  • [37] EDSD: efficient driving scenes detection based on Swin Transformer
    Wei Chen
    Ruihan Zheng
    Jiade Jiang
    Zijian Tian
    Fan Zhang
    Yi Liu
    Multimedia Tools and Applications, 2024, 83 (39) : 87179 - 87198
  • [38] Plant Disease Detection Algorithm Based on Efficient Swin Transformer
    Liu, Wei
    Zhang, Ao
    CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 82 (02): : 3045 - 3068
  • [39] Attention Swin Transformer UNet for Landslide Segmentation in Remotely Sensed Images
    Liu, Bingxue
    Wang, Wei
    Wu, Yuming
    Gao, Xing
    REMOTE SENSING, 2024, 16 (23)
  • [40] Efficient Approach to Detect and Localize Text in Natural Scene Images
    Samuel, S. R. Surem
    Christopher, C. Seldev
    ARTIFICIAL INTELLIGENCE AND EVOLUTIONARY ALGORITHMS IN ENGINEERING SYSTEMS, VOL 2, 2015, 325 : 665 - 674