Clothing generation by multi-modal embedding: A compatibility matrix-regularized GAN model

被引:6
|
作者
Liu, Linlin [1 ]
Zhang, Haijun [1 ]
Zhou, Dongliang [1 ]
机构
[1] Harbin Inst Technol, Dept Comp Sci, Shenzhen 518055, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Multi-modal embedding; Compatibility learning; Generative adversarial network; Image translation; Fashion data; IMAGE SYNTHESIS;
D O I
10.1016/j.imavis.2021.104097
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Clothing compatibility learning has gained increasing research attention due to the fact that a properly coordinated outfit can represent personality and improve an individual's appearance greatly. In this paper, we propose a Compatibility Matrix-Regularized Generative Adversarial Network (CMRGAN) for compatible item generation. In particular, we utilize a multi-modal embedding to transform the image and text information of an input clothing item into a latent feature code. Sequentially, compatibility learning among latent features is performed to obtain a compatibility style space. The feature of the input image is then regularized by the style space. Finally, a compatible clothing image is generated by a decoder which is fed by the regularized features. To verify the proposed model, we train an Inception-v3 classification model to evaluate the authenticity of synthesized images, a regression scoring VGG model to measure the compatibility degree of the generated image pairs and a deep attentional multimodal similarity model to evaluate the semantic similarity between generated images and ground truth text descriptions. In order to give an objective evaluation, these models are trained based on datasets consisting of fashion data only. The results demonstrate the effectiveness of the proposed method on image to-image translation based on compatibility space. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] MGCM: Multi-modal generative compatibility modeling for clothing matching
    Liu, Jinhuan
    Song, Xuemeng
    Chen, Zhumin
    Ma, Jun
    [J]. NEUROCOMPUTING, 2020, 414 : 215 - 224
  • [2] MCCP: multi-modal fashion compatibility and conditional preference model for personalized clothing recommendation
    Wang, Yunzhu
    Liu, Li
    Fu, Xiaodong
    Liu, Lijun
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 9621 - 9645
  • [3] MCCP: multi-modal fashion compatibility and conditional preference model for personalized clothing recommendation
    Yunzhu Wang
    Li Liu
    Xiaodong Fu
    Lijun Liu
    [J]. Multimedia Tools and Applications, 2024, 83 : 9621 - 9645
  • [4] ERDBF: Embedding-Regularized Double Branches Fusion for Multi-Modal Age Estimation
    Wu, Bo
    Lu, Hengjie
    Chen, Zhiyong
    Zhu, Congcong
    Xu, Shugong
    [J]. IEEE ACCESS, 2023, 11 : 47608 - 47618
  • [5] A Multi-Modal Chinese Poetry Generation Model
    Liu, Dayiheng
    Guo, Quan
    Li, Wubo
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [6] Lightweight multi-modal emotion recognition model based on modal generation
    Liu, Peisong
    Che, Manqiang
    Luo, Jiangchuan
    [J]. 2022 9TH INTERNATIONAL FORUM ON ELECTRICAL ENGINEERING AND AUTOMATION, IFEEA, 2022, : 430 - 435
  • [7] Hadamard matrix-guided multi-modal hashing for multi-modal retrieval
    Yu, Jun
    Huang, Wei
    Li, Zuhe
    Shu, Zhenqiu
    Zhu, Liang
    [J]. DIGITAL SIGNAL PROCESSING, 2022, 130
  • [8] Unsupervised cross-modal retrieval via Multi-modal graph regularized Smooth Matrix Factorization Hashing
    Fang, Yixian
    Zhang, Huaxiang
    Ren, Yuwei
    [J]. KNOWLEDGE-BASED SYSTEMS, 2019, 171 : 69 - 80
  • [9] Co-Regularized PLSA for Multi-Modal Learning
    Wang, Xin
    Chang, Ming-Ching
    Ying, Yiming
    Lyu, Siwei
    [J]. THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2166 - 2172
  • [10] Supervised Multi-modal Dictionary Learning for Clothing Representation
    Zhao, Qilu
    Wang, Jiayan
    Li, Zongmin
    [J]. PROCEEDINGS OF THE FIFTEENTH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS - MVA2017, 2017, : 51 - 54