DeepPortraitDrawing: Generating human body images from freehand sketches

被引:4
|
作者
Wu, Xian [1 ]
Wang, Chen [2 ]
Fu, Hongbo [3 ]
Shamir, Ariel [4 ]
Zhang, Song-Hai [2 ]
机构
[1] Bytedance, Beijing 100028, Peoples R China
[2] Tsinghua Univ, Beijing 100084, Peoples R China
[3] City Univ Hong Kong, Hong Kong 518057, Peoples R China
[4] Reichman Univ, IL-4610101 Herzliyya, Israel
来源
COMPUTERS & GRAPHICS-UK | 2023年 / 116卷
关键词
Generative adversarial networks; Image-to-image generation; StyleGAN;
D O I
10.1016/j.cag.2023.08.005
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Various methods for generating realistic images of objects and human faces from freehand sketches have been explored. However, generating realistic human body images from sketches is still a challenging problem. It is, first because of the sensitivity to human shapes, second because of the complexity of human images caused by body shape and pose changes, and third because of the domain gap between realistic images and freehand sketches. In this work, we present DeepPortraitDrawing, a deep generative framework for converting roughly drawn sketches to realistic human body images. To encode complicated body shapes under various poses, we take a local-to-global approach. Locally, we employ semantic part auto-encoders to construct part-level shape spaces, which are useful for refining the geometry of an input pre-segmented hand-drawn sketch. Globally, we employ a cascaded spatial transformer network to refine the structure of body parts by adjusting their spatial locations and relative proportions. Finally, we use a style-based generator as the global synthesis network for the sketch-to-image translation task which is modulated by segmentation maps for semantic preservation. Extensive experiments have shown that given roughly sketched human portraits, our method produces more realistic images than the state-of-the-art sketch-to-image synthesis techniques.& COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:73 / 81
页数:9
相关论文
共 50 条
  • [41] CanonicalFusion: Generating Drivable 3D Human Avatars from Multiple Images
    Shin, Jisu
    Lee, Junmyeong
    Lee, Seongmin
    Park, Min-Gyu
    Kang, Ju-Mi
    Yoon, Ju Hong
    Jeon, Hae-Gon
    COMPUTER VISION - ECCV 2024, PT XXIII, 2025, 15081 : 38 - 56
  • [42] Method for Generating Synthetic Images of Masked Human Faces
    Letenkov M.A.
    Iakovlev R.N.
    Markitantov M.V.
    Ryumin D.A.
    Saveliev A.I.
    Karpov A.A.
    Scientific Visualization, 2022, 14 (02): : 1 - 17
  • [43] Generating summaries for photographic images based on human affects
    Kim, Eun Yi
    Ko, Eunjeong
    PROCEEDINGS OF 2015 IEEE 14TH INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS & COGNITIVE COMPUTING (ICCI*CC), 2015, : 360 - 367
  • [44] Direct surface extraction from 3D freehand ultrasound images
    Zhang, Y
    Rohling, R
    Pai, DK
    VIS 2002: IEEE VISUALIZATION 2002, PROCEEDINGS, 2002, : 45 - 52
  • [45] WAYLA - Generating Images from Eye Movements
    Yu, Bingqing
    Clark, James J.
    2018 15TH CONFERENCE ON COMPUTER AND ROBOT VISION (CRV), 2018, : 118 - 125
  • [46] Generating Segmented Quality Meshes from Images
    Cuadros-Vargas, A. J.
    Lizier, M.
    Minghim, R.
    Nonato, L. G.
    JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2009, 33 (01) : 11 - 23
  • [47] Generating text from functional brain images
    Pereira, Francisco
    Detre, Greg
    Botvinick, Matthew
    FRONTIERS IN HUMAN NEUROSCIENCE, 2011, 5
  • [48] Generating Virtual Images from Oblique Frames
    Tommaselli, Antonio M. G.
    Galo, Mauricio
    de Moraes, Marcus V. A.
    Marcato, Jose, Jr.
    Caldeira, Carlos R. T.
    Lopes, Rodrigo F.
    REMOTE SENSING, 2013, 5 (04): : 1875 - 1893
  • [49] Generating Segmented Quality Meshes from Images
    A. J. Cuadros-Vargas
    M. Lizier
    R. Minghim
    L. G. Nonato
    Journal of Mathematical Imaging and Vision, 2009, 33 : 11 - 23
  • [50] Human Body Pose Estimation from Still Images and Video Frames
    El-Sallam, Amar A.
    Mian, Ajmal S.
    IMAGE ANALYSIS AND RECOGNITION, PT I, PROCEEDINGS, 2010, 6111 : 176 - 188