PRE-TRAINING WITH FRACTAL IMAGES FACILITATES LEARNED IMAGE QUALITY ESTIMATION

被引:0
|
作者
Silbernagel, Malte [1 ]
Wiegand, Thomas [1 ]
Eisert, Peter [1 ]
Bosse, Sebastian [1 ]
机构
[1] Fraunhofer HHI, Berlin, Germany
关键词
neural networks; pre-training; image quality estimation; perception models; fractals;
D O I
10.1109/ICIP49359.2023.10222630
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Today's image quality estimation is widely dominated by learning-based approaches. The availability of annotated, i.e. rated, images is often a bottleneck in training datadriven visual quality models and hinders their generalization power. This paper proposed a novel pre-training scheme for learning-based quality estimation that does not rely on human-annotated datasets, but leverages synthetic fractal images. These images can be synthesized inexhaustibly and are inherently labeled during generation. We evaluate the pre-training strategy on a popular neural network-based quality model and show that the training effort can be reduced significantly, resulting in better final accuracy and faster convergence speed.
引用
收藏
页码:2625 / 2629
页数:5
相关论文
共 50 条
  • [1] Improving Fractal Pre-training
    Anderson, Connor
    Farrell, Ryan
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 2412 - 2421
  • [2] Pre-Training Without Natural Images
    Kataoka, Hirokatsu
    Okayasu, Kazushige
    Matsumoto, Asato
    Yamagata, Eisuke
    Yamada, Ryosuke
    Inoue, Nakamasa
    Nakamura, Akio
    Satoh, Yutaka
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (04) : 990 - 1007
  • [3] Pre-Training Without Natural Images
    Hirokatsu Kataoka
    Kazushige Okayasu
    Asato Matsumoto
    Eisuke Yamagata
    Ryosuke Yamada
    Nakamasa Inoue
    Akio Nakamura
    Yutaka Satoh
    [J]. International Journal of Computer Vision, 2022, 130 : 990 - 1007
  • [4] Image Representations Learned With Unsupervised Pre-Training Contain Human-like Biases
    Steed, Ryan
    Caliskan, Aylin
    [J]. PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, : 701 - 713
  • [5] Quality Diversity for Visual Pre-Training
    Chavhan, Ruchika
    Gouk, Henry
    Li, Da
    Hospedales, Timothy
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5361 - 5371
  • [6] In Defense of Image Pre-Training for Spatiotemporal Recognition
    Li, Xianhang
    Wang, Huiyu
    Wei, Chen
    Mei, Jieru
    Yuille, Alan
    Zhou, Yuyin
    Xie, Cihang
    [J]. COMPUTER VISION, ECCV 2022, PT XXV, 2022, 13685 : 675 - 691
  • [7] Dialogue-adaptive language model pre-training from quality estimation
    Li, Junlong
    Zhang, Zhuosheng
    Zhao, Hai
    [J]. NEUROCOMPUTING, 2023, 516 : 27 - 35
  • [8] Sigmoid Loss for Language Image Pre-Training
    Zhai, Xiaohua
    Mustafa, Basil
    Kolesnikov, Alexander
    Beyer, Lucas
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11941 - 11952
  • [9] Grounded Language-Image Pre-training
    Li, Liunian Harold
    Zhang, Pengchuan
    Zhang, Haotian
    Yang, Jianwei
    Li, Chunyuan
    Zhong, Yiwu
    Wang, Lijuan
    Yuan, Lu
    Zhang, Lei
    Hwang, Jenq-Neng
    Chang, Kai-Wei
    Gao, Jianfeng
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 10955 - 10965
  • [10] Symbolizing Visual Features for Pre-training with Unlabeled Images
    Kamata, Yuichi
    Yamada, Moyuru
    Kato, Keizo
    Nakagawa, Akira
    Okatani, Takayuki
    [J]. PATTERN RECOGNITION, ACPR 2021, PT II, 2022, 13189 : 490 - 503