SPIQ: A Self-Supervised Pre-Trained Model for Image Quality Assessment

被引:17
|
作者
Chen, Pengfei [1 ]
Li, Leida [2 ]
Wu, Qingbo [3 ]
Wu, Jinjian [2 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Jiangsu, Peoples R China
[2] Xidian Univ, Sch Artificial Intelligence, Xian 710071, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
关键词
Distortion; Feature extraction; Task analysis; Transformers; Training; Predictive models; Image quality; Blind image quality assessment; self-supervised pre-training; contrastive learning; INDEX;
D O I
10.1109/LSP.2022.3145326
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Blind image quality assessment (BIQA) has witnessed a flourishing progress due to the rapid advances in deep learning technique. The vast majority of prior BIQA methods try to leverage models pre-trained on ImageNet to mitigate the data shortage problem. These well-trained models, however, can be sub-optimal when applied to BIQA task that varies considerably from the image classification domain. To address this issue, we make the first attempt to leverage the plentiful unlabeled data to conduct self-supervised pre-training for BIQA task. Based on the distorted images generated from the high-quality samples using the designed distortion augmentation strategy, the proposed pre-training is implemented by a feature representation prediction task. Specifically, patch-wise feature representations corresponding to a certain grid are integrated to make prediction for the representation of the patch below it. The prediction quality is then evaluated using a contrastive loss to capture quality-aware information for BIQA task. Experimental results conducted on KADID-10 k and KonIQ-10 k databases demonstrate that the learned pre-trained model can significantly benefit the existing learning based IQA models.
引用
收藏
页码:513 / 517
页数:5
相关论文
共 50 条
  • [41] Image quality assessment based on self-supervised learning and knowledge distillation
    Sang, Qingbing
    Shu, Ziru
    Liu, Lixiong
    Hu, Cong
    Wu, Qin
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 90
  • [42] Toward Leveraging Pre-Trained Self-Supervised Frontends for Automatic Singing Voice Understanding Tasks: Three Case Studies
    Yamamoto, Yuya
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 1745 - 1752
  • [43] COMPARATIVE ANALYSIS OF SELF-SUPERVISED PRE-TRAINED VISION TRANSFORMERS AND CONVOLUTIONAL NEURAL NETWORKS WITH CHEXNET IN CLASSIFYING LUNG CONDITIONS
    Elwirehardja, Gregorius natanael
    Liem, Steve marcello
    Adjie, Maria linneke
    Tjan, Farrel alexander
    Setiawan, Joselyn
    Syahputra, Muhammad edo
    Muljo, Hery harjono
    COMMUNICATIONS IN MATHEMATICAL BIOLOGY AND NEUROSCIENCE, 2025,
  • [44] ASiT-CRNN: A method for sound event detection with fine-tuning of self-supervised pre-trained ASiT-based model
    Zheng, Yueyang
    Zhang, Ruikun
    Atito, Sara
    Yang, Shuguo
    Wang, Wenwu
    Mei, Yiduo
    DIGITAL SIGNAL PROCESSING, 2025, 160
  • [45] Applying Self-Supervised Learning to Image Quality Assessment in Chest CT Imaging
    Pouget, Eleonore
    Dedieu, Veronique
    BIOENGINEERING-BASEL, 2024, 11 (04):
  • [46] Improved Image Quality Assessment by Utilizing Pre-Trained Architecture Features with Unified Learning Mechanism
    Ryu, Jihyoung
    APPLIED SCIENCES-BASEL, 2023, 13 (04):
  • [47] Pre-Trained Image Processing Transformer
    Chen, Hanting
    Wang, Yunhe
    Guo, Tianyu
    Xu, Chang
    Deng, Yiping
    Liu, Zhenhua
    Ma, Siwei
    Xu, Chunjing
    Xu, Chao
    Gao, Wen
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 12294 - 12305
  • [48] Targeted Image Reconstruction by Sampling Pre-trained Diffusion Model
    Zheng, Jiageng
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, INTELLISYS 2023, 2024, 822 : 552 - 560
  • [49] Grading the severity of diabetic retinopathy using an ensemble of self-supervised pre-trained convolutional neural networks: ESSP-CNNs
    Parsa S.
    Khatibi T.
    Multimedia Tools and Applications, 2024, 83 (42) : 89837 - 89870
  • [50] Diabetic Retinopathy Classification with pre-trained Image Enhancement Model
    Mudaser, Wahidullah
    Padungweang, Praisan
    Mongkolnam, Pornchai
    Lavangnananda, Patcharaporn
    2021 IEEE 12TH ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), 2021, : 629 - 632