Variational Transformer Networks for Layout Generation

被引:49
|
作者
Arroyo, Diego Martin [1 ]
Postels, Janis [2 ]
Tombari, Federico [1 ,3 ]
机构
[1] Google Inc, Mountain View, CA 94043 USA
[2] Swiss Fed Inst Technol, Zurich, Switzerland
[3] Tech Univ Munich, Munich, Germany
关键词
D O I
10.1109/CVPR46437.2021.01343
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generative models able to synthesize layouts of different kinds (e.g. documents, user interfaces or furniture arrangements) are a useful tool to aid design processes and as a first step in the generation of synthetic data, among other tasks. We exploit the properties of self-attention layers to capture high level relationships between elements in a layout, and use these as the building blocks of the well-known Variational Autoencoder (VAE) formulation. Our proposed Variational Transformer Network (VTN) is capable of learning margins, alignments and other global design rules without explicit supervision. Layouts sampled from our model have a high degree of resemblance to the training data, while demonstrating appealing diversity. In an extensive evaluation on publicly available benchmarks for different layout types VTNs achieve state-of-the-art diversity and perceptual quality. Additionally, we show the capabilities of this method as part of a document layout detection pipeline.
引用
收藏
页码:13637 / 13647
页数:11
相关论文
共 50 条
  • [1] Geometry Aligned Variational Transformer for Image-conditioned Layout Generation
    Cao, Yunning
    Ma, Ye
    Zhou, Min
    Liu, Chuanbin
    Xie, Hongtao
    Ge, Tiezheng
    Jiang, Yuning
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1561 - 1571
  • [2] Self-refined variational transformer for image-conditioned layout generation
    Cao, Yunning
    Liu, Chuanbin
    Ma, Ye
    Zhou, Min
    Ge, Tiezheng
    Jiang, Yuning
    Xie, Hongtao
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024,
  • [3] BLT: Bidirectional Layout Transformer for Controllable Layout Generation
    Kong, Xiang
    Jiang, Lu
    Chang, Huiwen
    Zhang, Han
    Hao, Yuan
    Gong, Haifeng
    Essa, Irfan
    [J]. COMPUTER VISION - ECCV 2022, PT XVII, 2022, 13677 : 474 - 490
  • [4] WakaVT: A Sequential Variational Transformer for Waka Generation
    Yuka Takeishi
    Mingxuan Niu
    Jing Luo
    Zhong Jin
    Xinyu Yang
    [J]. Neural Processing Letters, 2022, 54 : 731 - 750
  • [5] WakaVT: A Sequential Variational Transformer for Waka Generation
    Takeishi, Yuka
    Niu, Mingxuan
    Luo, Jing
    Jin, Zhong
    Yang, Xinyu
    [J]. NEURAL PROCESSING LETTERS, 2022, 54 (02) : 731 - 750
  • [6] DLT: Conditioned layout generation with Joint Discrete-Continuous Diffusion Layout Transformer
    Levi, Elad
    Brosh, Eli
    Mykhailych, Mykola
    Perez, Meir
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 2106 - 2115
  • [7] A Transformer-Based Variational Autoencoder for Sentence Generation
    Liu, Danyang
    Liu, Gongshen
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [8] LayoutDM: Transformer-based Diffusion Model for Layout Generation
    Chai, Shang
    Zhuang, Liansheng
    Yan, Fengying
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18349 - 18358
  • [9] Patch layout generation by detecting feature networks
    Cao, Yuanhao
    Yan, Dong-Ming
    Wonka, Peter
    [J]. COMPUTERS & GRAPHICS-UK, 2015, 46 : 275 - 282
  • [10] Graph Transformer GANs With Graph Masked Modeling for Architectural Layout Generation
    Tang, Hao
    Shao, Ling
    Sebe, Nicu
    Van Gool, Luc
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (06) : 4298 - 4313