Face Inpainting with Pre-trained Image Transformers

被引:0
|
作者
Gonc, Kaan [1 ]
Saglam, Baturay [2 ]
Kozat, Suleyman S. [2 ]
Dibeklioglu, Hamdi [1 ]
机构
[1] Bilkent Univ, Bilgisayar Muhendisligi Bolumu, Ankara, Turkey
[2] Bilkent Univ, Elekt & Elekt Muhendisligi Bolumu, Ankara, Turkey
关键词
image inpainting; transformers; deep generative models;
D O I
10.1109/SIU55565.2022.9864676
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Image inpainting is an underdetermined inverse problem that allows various contents to fill in the missing or damaged regions realistically. Convolutional neural networks (CNNs) are commonly used to create aesthetically pleasing content, yet CNNs have restricted perception fields for collecting global characteristics. Transformers enable long-range relationships to be modeled and different content generated with autoregressive modeling of pixel-sequence distributions using image-level attention mechanism. However, the current approaches to inpainting with transformers are limited to task-specific datasets and require larger-scale data. We introduce an approach to image inpainting by leveraging pre-trained vision transformers to remedy this issue. Experiments show that our approach can outperform CNN-based approaches and have a remarkable performance closer to the task-specific transformer methods.
引用
收藏
页数:4
相关论文
共 50 条
  • [1] Are Pre-trained Convolutions Better than Pre-trained Transformers?
    Tay, Yi
    Dehghani, Mostafa
    Gupta, Jai
    Aribandi, Vamsi
    Bahri, Dara
    Qin, Zhen
    Metzler, Donald
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4349 - 4359
  • [2] Calibration of Pre-trained Transformers
    Desai, Shrey
    Durrett, Greg
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 295 - 302
  • [3] Emergent Modularity in Pre-trained Transformers
    Zhang, Zhengyan
    Zeng, Zhiyuan
    Lin, Yankai
    Xiao, Chaojun
    Wang, Xiaozhi
    Han, Xu
    Liu, Zhiyuan
    Xie, Ruobing
    Sun, Maosong
    Zhou, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4066 - 4083
  • [4] Pre-trained transformers: an empirical comparison
    Casola, Silvia
    Lauriola, Ivano
    Lavelli, Alberto
    MACHINE LEARNING WITH APPLICATIONS, 2022, 9
  • [5] ViTMatte: Boosting image matting with pre-trained plain vision transformers
    Yao, Jingfeng
    Wang, Xinggang
    Yang, Shusheng
    Wang, Baoyuan
    INFORMATION FUSION, 2024, 103
  • [6] How Different are Pre-trained Transformers for Text Ranking?
    Rau, David
    Kamps, Jaap
    ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 : 207 - 214
  • [7] Efficient feature selection for pre-trained vision transformers
    Huang, Lan
    Zeng, Jia
    Yu, Mengqiang
    Ding, Weiping
    Bai, Xingyu
    Wang, Kangping
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 254
  • [8] Predicting Terms in IS-A Relations with Pre-trained Transformers
    Nikishina, Irina
    Chernomorchenko, Polina
    Demidova, Anastasiia
    Panchenko, Alexander
    Biemann, Chris
    13TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING AND THE 3RD CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, IJCNLP-AACL 2023, 2023, : 134 - 148
  • [9] Generative pre-trained transformers (GPT) for surface engineering
    Kamnis, Spyros
    SURFACE & COATINGS TECHNOLOGY, 2023, 466
  • [10] Generating Extended and Multilingual Summaries with Pre-trained Transformers
    Calizzano, Remi
    Ostendorff, Malte
    Ruan, Qian
    Rehm, Georg
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 1640 - 1650