Ultra-high-definition underwater image enhancement via dual-domain interactive transformer network

被引:0
|
作者
Li, Weiwei [1 ]
Cao, Feiyuan [2 ,3 ]
Wei, Yiwen [2 ,3 ]
Shi, Zhenghao [4 ]
Jia, Xiuyi [2 ,3 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
[2] Guangxi Normal Univ, Guangxi Key Lab Multisource Informat Min & Secur, Guilin 541004, Guangxi, Peoples R China
[3] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[4] Xian Univ Technol, Sch Comp Sci & Engn, Xian 710048, Peoples R China
关键词
Dual-branch network; Feature interaction; Ultra-high-definition image; DESIGN;
D O I
10.1007/s13042-024-02379-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The proliferation of ultra-high-definition (UHD) imaging device is increasingly being used for underwater image acquisition. However, due to light scattering and underwater impurities, UHD underwater images often suffer from color deviations and edge blurriness. Many studies have attempted to enhance underwater images by integrating frequency domain and spatial domain information. Nonetheless, these approaches often interactively fuse dual-domain features only in the final fusion module, neglecting the complementary and guiding roles of frequency domain and spatial domain features. Additionally, the extraction of dual-domain features is independent of each other, which leads to the sharp advantages and disadvantages of the dual-domain features extracted by these methods. Consequently, these methods impose high demands on the feature fusion capabilities of the fusion module. But in order to handle UHD underwater images, the fusion modules in these methods often stack only a limited number of convolution and activation function operations. This limitation results in insufficient fusion capability, leading to defects in the restoration of edges and colors in the images. To address these issues, we develop a dual-domain interaction network for enhancing UHD underwater images. The network takes into account both frequency domain and spatial domain features to complement and guide each other's feature extraction patterns, and fully integrates the dual-domain features in the model to better recover image details and colors. Specifically, the network consists of a U-shaped structure, where each layer is composed of dual-domain interaction transformer blocks containing interactive multi-head attention and interactive simple gate feed-forward networks. The interactive multi-head attention captures local interaction features of frequency domain and spatial domain information using convolution operation, followed by multi-head attention operation to extract global information of the mixed features. The interactive simple gate feed-forward network further enhances the model's dual-domain interaction capability and cross-dimensional feature extraction ability, resulting in clearer edges and more realistic colors in the images. Experimental results demonstrate that the performance of our proposal in enhancing underwater images is significantly better than existing methods.
引用
收藏
页码:2093 / 2109
页数:17
相关论文
共 50 条
  • [41] DDRNet: Dual-Domain Refinement Network for Remote Sensing Image Semantic Segmentation
    Yang, Zhenhao
    Bi, Fukun
    Hou, Xinghai
    Zhou, Dehao
    Wang, Yanping
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 20177 - 20189
  • [42] LIGHTWEIGHT DUAL-DOMAIN NETWORK FOR REAL-TIME MEDICAL IMAGE SEGMENTATION
    Liu, Libo
    Fan, Xinxin
    Zhang, Xiaodong
    Hu, Qingmao
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 396 - 400
  • [43] Dual-Domain Neural Network for Sparse-View Photoacoustic Image Reconstruction
    Shen Kang
    Liu Songde
    Shi Junhui
    Tian Chao
    CHINESE JOURNAL OF LASERS-ZHONGGUO JIGUANG, 2022, 49 (05):
  • [44] DTT-NET: DUAL-DOMAIN TRANSLATION TRANSFORMER FOR SEMI-SUPERVISED IMAGE DERAINING
    Chen, Ze-Bin
    Wang, Yuan-Gen
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1621 - 1625
  • [45] Underwater Image Enhancement Network Based on Dual Layers Regression
    Jia, Huidi
    Xiao, Yeqing
    Wang, Qiang
    Chen, Xiai
    Han, Zhi
    Tang, Yandong
    ELECTRONICS, 2024, 13 (01)
  • [46] DAPNet: Dual Attention Probabilistic Network for Underwater Image Enhancement
    Li, Xueyong
    Yu, Rui
    Zhang, Weidong
    Lu, Huimin
    Zhao, Wenyi
    Hou, Guojia
    Liang, Zheng
    IEEE JOURNAL OF OCEANIC ENGINEERING, 2025, 50 (01) : 178 - 191
  • [47] Spatial-Frequency Dual-Domain Feature Fusion Network for Low-Light Remote Sensing Image Enhancement
    Yao, Zishu
    Fan, Guodong
    Fan, Jinfu
    Gan, Min
    Chen, C. L. Philip
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [48] DCT-net: Dual-domain cross-fusion transformer network for MRI reconstruction
    Wang, Bin
    Lian, Yusheng
    Xiong, Xingchuang
    Zhou, Han
    Liu, Zilong
    Zhou, Xiaohao
    MAGNETIC RESONANCE IMAGING, 2024, 107 : 69 - 79
  • [49] SwinGAN: A dual-domain Swin Transformer-based generative adversarial network for MRI reconstruction
    Zhao, Xiang
    Yang, Tiejun
    Li, Bingjie
    Zhang, Xin
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 153
  • [50] Viewers' Susceptibility to Image Blurs in Watching Ultra-high-definition TV Correlates with Their Dynamic Visual Acuity
    Emoto, Masaki
    Fang, Yu
    Shioiri, Satoshi
    ITE TRANSACTIONS ON MEDIA TECHNOLOGY AND APPLICATIONS, 2019, 7 (02): : 103 - 110