Graphmax for Text Generation

被引:0
|
作者
Liu, Bin [1 ]
Yin, Guosheng [2 ]
机构
[1] Southwestern Univ Finance & Econ, Ctr Stat Res, Sch Stat, Chengdu, Peoples R China
[2] Univ Hong Kong, Dept Stat & Actuarial Sci, Hong Kong, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In text generation, a large language model (LM) makes a choice of each new word based only on the former selection of its context using the softmax function. Nevertheless, the link statistics information of concurrent words based on a scene-specific corpus is valuable in choosing the next word, which can help to ensure the topic of the generated text to be aligned with the current task. To fully explore the co-occurrence information, we propose a graphmax function for task-specific text generation. Using the graph-based regularization, graphmax enables the final word choice to be determined by both the global knowledge from the LM and the local knowledge from the scene-specific corpus. The traditional softmax function is regularized with a graph total variation (GTV) term, which incorporates the local knowledge into the LM and encourages the model to consider the statistical relationships between words in a scene-specific corpus. The proposed graphmax is versatile and can be readily plugged into any large pre-trained LM for text generation and machine translation. Through extensive experiments, we demonstrate that the new GTV-based regularization can improve performances in various natural language processing (NLP) tasks in comparison with existing methods. Moreover, through human experiments, we observe that participants can easily distinguish the text generated by graphmax or softmax.
引用
收藏
页码:823 / 848
页数:26
相关论文
共 50 条
  • [41] YOUNG WRITERS USE OF TEXT STRUCTURE IN EXPOSITORY TEXT GENERATION
    ENGLERT, CS
    STEWART, SR
    HIEBERT, EH
    JOURNAL OF EDUCATIONAL PSYCHOLOGY, 1988, 80 (02) : 143 - 151
  • [42] Distractor Generation Through Text-to-Text Transformer Models
    de-Fitero-Dominguez, David
    Garcia-Lopez, Eva
    Garcia-Cabot, Antonio
    del-Hoyo-Gabaldon, Jesus-Angel
    Moreno-Cediel, Antonio
    IEEE ACCESS, 2024, 12 : 25580 - 25589
  • [43] Hybridization of Intelligent Solutions Architecture for Text Understanding and Text Generation
    Ivaschenko, Anton
    Krivosheev, Arkadiy
    Stolbova, Anastasia
    Golovnin, Oleg
    APPLIED SCIENCES-BASEL, 2021, 11 (11):
  • [44] Open Domain Event Text Generation
    Fu, Zihao
    Bing, Lidong
    Lam, Wai
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7748 - 7755
  • [45] Image Generation from Text and Segmentation
    Osugi, Masato
    Vargas, Danilo Vasconcellos
    2022 TENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS, CANDARW, 2022, : 206 - 211
  • [46] MULTILINGUAL TEXT GENERATION BY A MONOLINGUAL USER
    SOMERS, H
    JONES, D
    META, 1992, 37 (04) : 647 - 656
  • [47] Learning to Transfer Prompts for Text Generation
    Li, Junyi
    Tang, Tianyi
    Nie, Jian-Yun
    Wen, Ji-Rong
    Zhaol, Wayne Xin
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3506 - 3518
  • [48] A Deep Learning Approach for Text Generation
    Elmogy, Ahmed
    Mahmoud, Belal
    Saleh, Mohamed
    29TH INTERNATIONAL CONFERENCE ON COMPUTER THEORY AND APPLICATIONS (ICCTA 2019), 2019, : 102 - 106
  • [49] A Contrastive Framework for Neural Text Generation
    Su, Yixuan
    Lan, Tian
    Wang, Yan
    Yogatama, Dani
    Kong, Lingpeng
    Collier, Nigel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [50] Challenges and Opportunities in Text Generation Explainability
    Amara, Kenza
    Sevastjanow, Rita
    El-Assady, Mennatallah
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, PT I, XAI 2024, 2024, 2153 : 244 - 264