SPIQ: A Self-Supervised Pre-Trained Model for Image Quality Assessment

被引:17
|
作者
Chen, Pengfei [1 ]
Li, Leida [2 ]
Wu, Qingbo [3 ]
Wu, Jinjian [2 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Jiangsu, Peoples R China
[2] Xidian Univ, Sch Artificial Intelligence, Xian 710071, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
关键词
Distortion; Feature extraction; Task analysis; Transformers; Training; Predictive models; Image quality; Blind image quality assessment; self-supervised pre-training; contrastive learning; INDEX;
D O I
10.1109/LSP.2022.3145326
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Blind image quality assessment (BIQA) has witnessed a flourishing progress due to the rapid advances in deep learning technique. The vast majority of prior BIQA methods try to leverage models pre-trained on ImageNet to mitigate the data shortage problem. These well-trained models, however, can be sub-optimal when applied to BIQA task that varies considerably from the image classification domain. To address this issue, we make the first attempt to leverage the plentiful unlabeled data to conduct self-supervised pre-training for BIQA task. Based on the distorted images generated from the high-quality samples using the designed distortion augmentation strategy, the proposed pre-training is implemented by a feature representation prediction task. Specifically, patch-wise feature representations corresponding to a certain grid are integrated to make prediction for the representation of the patch below it. The prediction quality is then evaluated using a contrastive loss to capture quality-aware information for BIQA task. Experimental results conducted on KADID-10 k and KonIQ-10 k databases demonstrate that the learned pre-trained model can significantly benefit the existing learning based IQA models.
引用
收藏
页码:513 / 517
页数:5
相关论文
共 50 条
  • [21] Prediction of MASH features from liver biopsy images using a pre-trained self-supervised learning model
    Wang, Yang
    Vyawahare, Saurabh
    McNeil, Carson
    Loo, Jessica
    Robbins, Marc
    Goldenberg, Roman
    JOURNAL OF HEPATOLOGY, 2024, 80 : S592 - S592
  • [22] Investigation of self-supervised pre-trained models for classification of voice quality from speech and neck surface accelerometer signals
    Kadiri, Sudarsana Reddy
    Javanmardi, Farhad
    Alku, Paavo
    COMPUTER SPEECH AND LANGUAGE, 2023, 83
  • [23] ON THE USE OF SELF-SUPERVISED PRE-TRAINED ACOUSTIC AND LINGUISTIC FEATURES FOR CONTINUOUS SPEECH EMOTION RECOGNITION
    Macary, Manon
    Tahon, Marie
    Esteve, Yannick
    Rousseau, Anthony
    2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, : 373 - 380
  • [24] Quality-aware Pre-trained Models for Blind Image Quality Assessment
    Zhao, Kai
    Yuan, Kun
    Sun, Ming
    Li, Mading
    Wen, Xing
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 22302 - 22313
  • [25] Employing bimodal representations to predict DNA bendability within a self-supervised pre-trained framework
    Yang, Minghao
    Zhang, Shichen
    Zheng, Zhihang
    Zhang, Pengfei
    Liang, Yan
    Tang, Shaojun
    NUCLEIC ACIDS RESEARCH, 2024, 52 (06)
  • [26] GhostEncoder: Stealthy backdoor attacks with dynamic triggers to pre-trained encoders in self-supervised learning
    Wang, Qiannan
    Yin, Changchun
    Fang, Liming
    Liu, Zhe
    Wang, Run
    Lin, Chenhao
    COMPUTERS & SECURITY, 2024, 142
  • [27] Prediction of Protein Tertiary Structure Using Pre-Trained Self-Supervised Learning Based on Transformer
    Kurniawan, Alif
    Jatmiko, Wisnu
    Hertadi, Rukman
    Habibie, Novian
    2020 5TH INTERNATIONAL WORKSHOP ON BIG DATA AND INFORMATION SECURITY (IWBIS 2020), 2020, : 75 - 80
  • [28] Can Self-Supervised Neural Representations Pre-Trained on Human Speech distinguish Animal Callers?
    Sarkar, Eklavya
    Magimai-Doss, Mathew
    INTERSPEECH 2023, 2023, : 1189 - 1193
  • [29] ON FINE-TUNING PRE-TRAINED SPEECH MODELS WITH EMA-TARGET SELF-SUPERVISED LOSS
    Yang, Hejung
    Kang, Hong-Goo
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 6360 - 6364
  • [30] A Systematic Review of Transformer-Based Pre-Trained Language Models through Self-Supervised Learning
    Kotei, Evans
    Thirunavukarasu, Ramkumar
    INFORMATION, 2023, 14 (03)