A Fuzzy Training Framework for Controllable Sequence-to-Sequence Generation

被引:1
|
作者
Li, Jiajia [1 ]
Wang, Ping [2 ]
Li, Zuchao [2 ]
Liu, Xi [3 ]
Utiyama, Masao [4 ]
Sumita, Eiichiro [4 ]
Zhao, Hai [5 ]
Ai, Haojun [2 ]
机构
[1] Hankou Univ, Mus Sch, Wuhan 430212, Peoples R China
[2] Wuhan Univ, Wuhan 430072, Peoples R China
[3] Wuhan Conservatory Mus, Wuhan 430060, Peoples R China
[4] Natl Inst Informat & Commun Technol, Koganei, Tokyo 1848795, Japan
[5] Shanghai Jiao Tong Univ, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Artificial intelligence; Decoding; Machine translation; Training data; Music; Natural languages; Computational modeling; Time factors; Fuzzy systems; Task analysis; Music lyrics generation; controllable generation; music understanding; constrained decoding; fuzzy training;
D O I
10.1109/ACCESS.2022.3202010
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The generation of music lyrics by artificial intelligence (AI) is frequently modeled as a language-targeted sequence-to-sequence generation task. Formally, if we convert the melody into a word sequence, we can consider the lyrics generation task to be a machine translation task. Traditional machine translation tasks involve translating between cross-lingual word sequences, whereas music lyrics generation tasks involve translating between music and natural language word sequences. The theme or key words of the generated lyrics are usually limited to meet the needs of the users when they are generated. This requirement can be thought of as a restricted translation problem. In this paper, we propose a fuzzy training framework that allows a model to simultaneously support both unrestricted and restricted translation by adopting an additional auxiliary training process without constraining the decoding process. This maintains the benefits of restricted translation but greatly reduces the extra time overhead of constrained decoding, thus improving its practicality. The experimental results show that our framework is well suited to the Chinese lyrics generation and restricted machine translation tasks, and that it can also generate language sequence under the condition of given restricted words without training multiple models, thereby achieving the goal of green AI.
引用
收藏
页码:92467 / 92480
页数:14
相关论文
共 50 条
  • [1] Controllable Question Generation via Sequence-to-Sequence Neural Model with Auxiliary Information
    Cao, Zhen
    Tatinati, Sivanagaraja
    Khong, Andy W. H.
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [2] Attention based sequence-to-sequence framework for auto image caption generation
    Khan, Rashid
    Islam, M. Shujah
    Kanwal, Khadija
    Iqbal, Mansoor
    Hossain, Md Imran
    Ye, Zhongfu
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 43 (01) : 159 - 170
  • [3] Denoising based Sequence-to-Sequence Pre-training for Text Generation
    Wang, Liang
    Zhao, Wei
    Jia, Ruoyu
    Li, Sujian
    Liu, Jingming
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 4003 - 4015
  • [4] Synthesizing waveform sequence-to-sequence to augment training data for sequence-to-sequence speech recognition
    Ueno, Sei
    Mimura, Masato
    Sakai, Shinsuke
    Kawahara, Tatsuya
    [J]. ACOUSTICAL SCIENCE AND TECHNOLOGY, 2021, 42 (06) : 333 - 343
  • [5] Data generation using sequence-to-sequence
    Joshi, Akshat
    Mehta, Kinal
    Gupta, Neha
    Valloli, Varun Kannadi
    [J]. 2018 IEEE RECENT ADVANCES IN INTELLIGENT COMPUTATIONAL SYSTEMS (RAICS), 2018, : 108 - 112
  • [6] Neural AMR: Sequence-to-Sequence Models for Parsing and Generation
    Konstas, Ioannis
    Iyer, Srinivasan
    Yatskar, Mark
    Choi, Yejin
    Zettlemoyer, Luke
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 146 - 157
  • [7] Myanmar News Headline Generation with Sequence-to-Sequence model
    Thu, Yamin
    Pa, Win Pa
    [J]. PROCEEDINGS OF 2020 23RD CONFERENCE OF THE ORIENTAL COCOSDA INTERNATIONAL COMMITTEE FOR THE CO-ORDINATION AND STANDARDISATION OF SPEECH DATABASES AND ASSESSMENT TECHNIQUES (ORIENTAL-COCOSDA 2020), 2020, : 117 - 122
  • [8] Encoding Emotional Information for Sequence-to-Sequence Response Generation
    Chan, Yin Hei
    Lui, Andrew Kwok Fai
    [J]. 2018 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD), 2018, : 113 - 116
  • [9] Persian Keyphrase Generation Using Sequence-to-sequence Models
    Doostmohammadi, Ehsan
    Bokaei, Mohammad Hadi
    Sameti, Hossein
    [J]. 2019 27TH IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE 2019), 2019, : 2010 - 2015
  • [10] Improving Sequence-to-Sequence Pre-training via Sequence Span Rewriting
    Zhou, Wangchunshu
    Ge, Tao
    Xu, Canwen
    Xu, Ke
    Wei, Furu
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 571 - 582