Copy Mechanism and Tailored Training for Character-Based Data-to-Text Generation

被引:1
|
作者
Roberti, Marco [1 ]
Bonetta, Giovanni [1 ]
Cancelliere, Rossella [1 ]
Gallinari, Patrick [2 ,3 ]
机构
[1] Univ Turin, Comp Sci Dept, Via Pessinetto 12, I-12149 Turin, Italy
[2] Sorbonne Univ, 4 Pl Jussieu, F-75005 Paris, France
[3] Criteo AI Lab, 32 Rue Blanche, F-75009 Paris, France
关键词
Natural language processing; Data-to-text generation; Deep learning; Sequence-to-sequence; Dataset;
D O I
10.1007/978-3-030-46147-8_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the last few years, many different methods have been focusing on using deep recurrent neural networks for natural language generation. The most widely used sequence-to-sequence neural methods are word-based: as such, they need a pre-processing step called delexicalization (conversely, relexicalization) to deal with uncommon or unknown words. These forms of processing, however, give rise to models that depend on the vocabulary used and are not completely neural. In this work, we present an end-to-end sequence-to-sequence model with attention mechanism which reads and generates at a character level, no longer requiring delexicalization, tokenization, nor even lowercasing. Moreover, since characters constitute the common "building blocks" of every text, it also allows a more general approach to text generation, enabling the possibility to exploit transfer learning for training. These skills are obtained thanks to two major features: (i) the possibility to alternate between the standard generation mechanism and a copy one, which allows to directly copy input facts to produce outputs, and (ii) the use of an original training pipeline that further improves the quality of the generated texts. We also introduce a new dataset called E2E+, designed to highlight the copying capabilities of character-based models, that is a modified version of the well-known E2E dataset used in the E2E Challenge. We tested our model according to five broadly accepted metrics (including the widely used bleu), showing that it yields competitive performance with respect to both character-based and word-based approaches.
引用
收藏
页码:648 / 664
页数:17
相关论文
共 50 条
  • [31] Controlling hallucinations at word level in data-to-text generation
    Clement Rebuffel
    Marco Roberti
    Laure Soulier
    Geoffrey Scoutheeten
    Rossella Cancelliere
    Patrick Gallinari
    Data Mining and Knowledge Discovery, 2022, 36 : 318 - 354
  • [32] A Data-to-Text Generation Model with Deduplicated Content Planning
    Wang, Mengda
    Cao, Jianjun
    Yu, Xu
    Nie, Zibo
    BIG DATA, BIGDATA 2022, 2022, 1709 : 92 - 103
  • [33] Neural Data-to-Text Generation Guided by Predicted Plan
    Gao, Hanning
    Wei, Zhihua
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON INFORMATION COMMUNICATION AND SOFTWARE ENGINEERING (ICICSE 2022), 2022, : 53 - 59
  • [34] Controlling hallucinations at word level in data-to-text generation
    Rebuffel, Clement
    Roberti, Marco
    Soulier, Laure
    Scoutheeten, Geoffrey
    Cancelliere, Rossella
    Gallinari, Patrick
    DATA MINING AND KNOWLEDGE DISCOVERY, 2022, 36 (01) : 318 - 354
  • [35] Unifying Structured Data as Graph for Data-to-Text Pre-Training
    Li, Shujie
    Li, Liang
    Geng, Ruiying
    Yang, Min
    Li, Binhua
    Yuan, Guanghu
    He, Wanwei
    Yuan, Shao
    Ma, Can
    Huang, Fei
    Li, Yongbin
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 210 - 228
  • [36] Partially-Aligned Data-to-Text Generation with Distant Supervision
    Fu, Zihao
    Shi, Bei
    Lam, Wai
    Bing, Lidong
    Liu, Zhiyuan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9183 - 9193
  • [37] Does the Order of Training Samples Matter? Improving Neural Data-to-Text Generation with Curriculum Learning
    Chang, Ernie
    Yeh, Hui-Syuan
    Demberg, Vera
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 727 - 733
  • [38] Character-based Search with Data Confidentiality in the Clouds
    Maheshwari, Varun
    Nourian, Arash
    Maheswaran, Muthucumaru
    2012 IEEE 4TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING TECHNOLOGY AND SCIENCE (CLOUDCOM), 2012,
  • [39] A Logic Aware Neural Generation Method for Explainable Data-to-text
    Lin, Xiexiong
    Li, Huaisong
    Huang, Tao
    Wang, Feng
    Chao, Linlin
    Zhuang, Fuzhen
    Wang, Taifeng
    Zhang, Tianyi
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 3318 - 3326
  • [40] Stochastic Data-to-Text Generation Using Syntactic Dependency Information
    Seifossadat, Elham
    Sameti, Hossein
    COMPUTER SPEECH AND LANGUAGE, 2022, 76