Graphmax for Text Generation

被引:0
|
作者
Liu, Bin [1 ]
Yin, Guosheng [2 ]
机构
[1] Southwestern Univ Finance & Econ, Ctr Stat Res, Sch Stat, Chengdu, Peoples R China
[2] Univ Hong Kong, Dept Stat & Actuarial Sci, Hong Kong, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In text generation, a large language model (LM) makes a choice of each new word based only on the former selection of its context using the softmax function. Nevertheless, the link statistics information of concurrent words based on a scene-specific corpus is valuable in choosing the next word, which can help to ensure the topic of the generated text to be aligned with the current task. To fully explore the co-occurrence information, we propose a graphmax function for task-specific text generation. Using the graph-based regularization, graphmax enables the final word choice to be determined by both the global knowledge from the LM and the local knowledge from the scene-specific corpus. The traditional softmax function is regularized with a graph total variation (GTV) term, which incorporates the local knowledge into the LM and encourages the model to consider the statistical relationships between words in a scene-specific corpus. The proposed graphmax is versatile and can be readily plugged into any large pre-trained LM for text generation and machine translation. Through extensive experiments, we demonstrate that the new GTV-based regularization can improve performances in various natural language processing (NLP) tasks in comparison with existing methods. Moreover, through human experiments, we observe that participants can easily distinguish the text generated by graphmax or softmax.
引用
收藏
页码:823 / 848
页数:26
相关论文
共 50 条
  • [1] Graphmax for Text Generation
    Liu B.
    Yin G.
    Journal of Artificial Intelligence Research, 2023, 78 : 823 - 848
  • [2] TEXT GENERATION
    KUNST, AE
    SUB-STANCE, 1977, (16) : 159 - 171
  • [3] Text Generation for Imbalanced Text Classification
    Akkaradamrongrat, Suphamongkol
    Kachamas, Pornpimon
    Sinthupinyo, Sukree
    2019 16TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING (JCSSE 2019), 2019, : 181 - 186
  • [4] Evaluating Text-to-Visual Generation with Image-to-Text Generation
    Lin, Zhiqiu
    Athaki, Deepak
    Li, Baiqi
    Li, Jiayao
    Xia, Xide
    Neubig, Graham
    Zhang, Pengchuan
    Ramanan, Deva
    COMPUTER VISION - ECCV 2024, PT IX, 2025, 15067 : 366 - 384
  • [5] THE NEED FOR TEXT GENERATION
    MCKEOWN, KR
    AFIPS CONFERENCE PROCEEDINGS, 1985, 54 : 87 - &
  • [6] Sparse Text Generation
    Martins, Pedro Henrique
    Marinho, Zita
    Martins, Andre F. T.
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 4252 - 4273
  • [7] AUTOMATIC TEXT GENERATION
    MARSH, R
    BRITISH TELECOM TECHNOLOGY JOURNAL, 1988, 6 (04): : 84 - 88
  • [8] Interactive Text Generation
    Faltings, Felix
    Galley, Michel
    Brantley, Kiante
    Peng, Baolin
    Cai, Weixin
    Zhang, Yizhe
    Gao, Jianfeng
    Dolan, Bill
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 4450 - 4468
  • [9] Stochastic text generation
    Oberlander, J
    Brew, C
    PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY OF LONDON SERIES A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2000, 358 (1769): : 1373 - 1386
  • [10] BARTSCORE: Evaluating Generated Text as Text Generation
    Yuan, Weizhe
    Neubig, Graham
    Liu, Pengfei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34