Infrared Image Generation Based on Visual State Space and Contrastive Learning

被引:0
|
作者
Li, Bing [1 ]
Ma, Decao [1 ]
He, Fang [1 ]
Zhang, Zhili [1 ]
Zhang, Daqiao [1 ]
Li, Shaopeng [1 ,2 ]
机构
[1] Xian Res Inst High Technol, Xian 710025, Peoples R China
[2] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
visual state space; contrastive learning; generative adversarial network; visible-to-infrared image translation;
D O I
10.3390/rs16203817
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
The preparation of infrared reference images is of great significance for improving the accuracy and precision of infrared imaging guidance. However, collecting infrared data on-site is difficult and time-consuming. Fortunately, the infrared images can be obtained from the corresponding visible-light images to enrich the infrared data. To this end, this present work proposes an image translation algorithm that converts visible-light images to infrared images. This algorithm, named V2IGAN, is founded on the visual state space attention module and multi-scale feature contrastive learning loss. Firstly, we introduce a visual state space attention module designed to sharpen the generative network's focus on critical regions within visible-light images. This enhancement not only improves feature extraction but also bolsters the generator's capacity to accurately model features, ultimately enhancing the quality of generated images. Furthermore, the method incorporates a multi-scale feature contrastive learning loss function, which serves to bolster the robustness of the model and refine the detail of the generated images. Experimental results show that the V2IGAN method outperforms existing typical infrared image generation techniques in both subjective visual assessments and objective metric evaluations. This suggests that the V2IGAN method is adept at enhancing the feature representation in images, refining the details of the generated infrared images, and yielding reliable, high-quality results.
引用
收藏
页数:24
相关论文
共 50 条
  • [21] CLF-Net: Contrastive Learning for Infrared and Visible Image Fusion Network
    Zhu, Zhengjie
    Yang, Xiaogang
    Lu, Ruitao
    Shen, Tong
    Xie, Xueli
    Zhang, Tao
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [22] Image space trajectory generation for image-based visual servoing under large pose error
    Park, JS
    Chung, MJ
    IROS 2001: PROCEEDINGS OF THE 2001 IEEE/RJS INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-4: EXPANDING THE SOCIETAL ROLE OF ROBOTICS IN THE NEXT MILLENNIUM, 2001, : 1159 - 1164
  • [23] Few-shot image generation based on contrastive meta-learning generative adversarial network
    Aniwat Phaphuangwittayakul
    Fangli Ying
    Yi Guo
    Liting Zhou
    Nopasit Chakpitak
    The Visual Computer, 2023, 39 : 4015 - 4028
  • [24] Few鄄Shot Image Classification Based on Local Contrastive Learning and Novel Class Feature Generation
    Chen, Ning
    Liu, Fan
    Dong, Chenwei
    Chen, Zhiyu
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2024, 37 (10): : 936 - 946
  • [25] Vi2CLR: Video and Image for Visual Contrastive Learning of Representation
    Diba, Ali
    Sharma, Vivek
    Safdari, Reza
    Lotfi, Dariush
    Sarfraz, M. Saquib
    Stiefelhagen, Rainer
    Van Gool, Luc
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1482 - 1492
  • [26] Colorization of infrared images based on feature fusion and contrastive learning
    Chen, Lingqiang
    Liu, Yuan
    He, Yin
    Xie, Zhihua
    Sui, Xiubao
    OPTICS AND LASERS IN ENGINEERING, 2023, 162
  • [27] GAN-Based Inter-Class Sample Generation for Contrastive Learning of Vein Image Representations
    Ou, Wei-Feng
    Po, Lai-Man
    Zhou, Chang
    Xian, Peng-Fei
    Xiong, Jing-Jing
    IEEE TRANSACTIONS ON BIOMETRICS, BEHAVIOR, AND IDENTITY SCIENCE, 2022, 4 (02): : 249 - 262
  • [28] Few-shot image generation based on contrastive meta-learning generative adversarial network
    Phaphuangwittayakul, Aniwat
    Ying, Fangli
    Guo, Yi
    Zhou, Liting
    Chakpitak, Nopasit
    VISUAL COMPUTER, 2023, 39 (09): : 4015 - 4028
  • [29] Contrastive Masked Image-Text Modeling for Medical Visual Representation Learning
    Chen, Cheng
    Zhong, Aoxiao
    Wu, Dufan
    Luo, Jie
    Li, Quanzheng
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT V, 2023, 14224 : 493 - 503
  • [30] Regularizing Visual Semantic Embedding With Contrastive Learning for Image-Text Matching
    Liu, Yang
    Liu, Hong
    Wang, Huaqiu
    Liu, Mengyuan
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1332 - 1336