An automatic generation method of cross-modal fuzzy creativity

被引:2
|
作者
Zhang, Fuquan [1 ]
Wang, Yiou [2 ]
Wu, Chensheng [2 ]
机构
[1] Minjiang Univ, Fujian Prov Key Lab Informat Proc & Intelligent C, Fuzhou, Peoples R China
[2] Beijing Inst Sci & Technol Informat, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Generation of fuzzy creativity; cross-modal; graph neural network; creative works;
D O I
10.3233/JIFS-179657
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Digital creativity is creative expression derived from cultural creativity and information technology. In order to overcome the problem in the creative generation in the condition of fuzzy and uncertain ideas, an automatic generation method of cross-modal fuzzy creativity (AGMCFC) is proposed. In this subject, fuzzy creative data sets and learning retrieval network are constructed for the sake of extracting original creative data effectively. And the logical correlations between creative objects are acquired dynamically based on the graph neural network. Creative objects and creative styles are generated by using generative adversarial nets technology and style transfer technology, respectively. Then, the projectiles, boundary markers and location words of the creative scene objects are generated by analyzing related attributes of each entity. After adjusting the layout, creative works are automatically generated. A fuzzy creative generating environment is implemented. Experimental results show that the screened number of AGMCFC method is about twice as much as that of manual method, and the accuracy rate of AGMCFC method is improved compared with the manual method. AGMCFC method performs well at creative generation of fuzzy ideas automatically.
引用
收藏
页码:5685 / 5696
页数:12
相关论文
共 50 条
  • [1] Cross-Modal Commentator: Automatic Machine Commenting Based on Cross-Modal Information
    Yang, Pengcheng
    Zhang, Zhihan
    Luo, Fuli
    Li, Lei
    Huang, Chengyang
    Sun, Xu
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2680 - 2686
  • [2] Cross-Modal Knowledge Distillation Method for Automatic Cued Speech Recognition
    Wang, Jianrong
    Tang, Ziyue
    Li, Xuewei
    Yu, Mei
    Fang, Qiang
    Liu, Li
    [J]. INTERSPEECH 2021, 2021, : 2986 - 2990
  • [3] A cross-modal crowd counting method combining CNN and cross-modal transformer
    Zhang, Shihui
    Wang, Wei
    Zhao, Weibo
    Wang, Lei
    Li, Qunpeng
    [J]. IMAGE AND VISION COMPUTING, 2023, 129
  • [4] Creativity and Perception: Unveiling the Role of Cross-Modal Audiovisual Integration
    Xiao, Xinyao
    Wang, Jian
    Shu, Yanyan
    Tan, Junying
    [J]. JOURNAL OF CREATIVE BEHAVIOR, 2024,
  • [5] A Survey of Cross-Modal Visual Content Generation
    Nazarieh, Fatemeh
    Feng, Zhenhua
    Awais, Muhammad
    Wang, Wenwu
    Kittler, Josef
    [J]. IEEE Transactions on Circuits and Systems for Video Technology, 2024, 34 (08) : 6814 - 6832
  • [6] Audio-to-Image Cross-Modal Generation
    Zelaszczyk, Maciej
    Mandziuk, Jacek
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [7] Deep Cross-Modal Audio-Visual Generation
    Chen, Lele
    Srivastava, Sudhanshu
    Duan, Zhiyao
    Xu, Chenliang
    [J]. PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, : 349 - 357
  • [8] Reinforced Cross-modal Alignment for Radiology Report Generation
    Qin, Han
    Song, Yan
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 448 - 458
  • [9] A cross-modal method of labeling music tags
    Hsu, Jia-Lien
    Li, Yen-Fu
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2012, 58 (03) : 521 - 541
  • [10] A cross-modal method of labeling music tags
    Jia-Lien Hsu
    Yen-Fu Li
    [J]. Multimedia Tools and Applications, 2012, 58 : 521 - 541