Copy Mechanism and Tailored Training for Character-Based Data-to-Text Generation

被引:2
|
作者
Roberti, Marco [1 ]
Bonetta, Giovanni [1 ]
Cancelliere, Rossella [1 ]
Gallinari, Patrick [2 ,3 ]
机构
[1] Univ Turin, Comp Sci Dept, Via Pessinetto 12, I-12149 Turin, Italy
[2] Sorbonne Univ, 4 Pl Jussieu, F-75005 Paris, France
[3] Criteo AI Lab, 32 Rue Blanche, F-75009 Paris, France
关键词
Natural language processing; Data-to-text generation; Deep learning; Sequence-to-sequence; Dataset;
D O I
10.1007/978-3-030-46147-8_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the last few years, many different methods have been focusing on using deep recurrent neural networks for natural language generation. The most widely used sequence-to-sequence neural methods are word-based: as such, they need a pre-processing step called delexicalization (conversely, relexicalization) to deal with uncommon or unknown words. These forms of processing, however, give rise to models that depend on the vocabulary used and are not completely neural. In this work, we present an end-to-end sequence-to-sequence model with attention mechanism which reads and generates at a character level, no longer requiring delexicalization, tokenization, nor even lowercasing. Moreover, since characters constitute the common "building blocks" of every text, it also allows a more general approach to text generation, enabling the possibility to exploit transfer learning for training. These skills are obtained thanks to two major features: (i) the possibility to alternate between the standard generation mechanism and a copy one, which allows to directly copy input facts to produce outputs, and (ii) the use of an original training pipeline that further improves the quality of the generated texts. We also introduce a new dataset called E2E+, designed to highlight the copying capabilities of character-based models, that is a modified version of the well-known E2E dataset used in the E2E Challenge. We tested our model according to five broadly accepted metrics (including the widely used bleu), showing that it yields competitive performance with respect to both character-based and word-based approaches.
引用
收藏
页码:648 / 664
页数:17
相关论文
共 50 条
  • [41] Stochastic Data-to-Text Generation Using Syntactic Dependency Information
    Seifossadat, Elham
    Sameti, Hossein
    COMPUTER SPEECH AND LANGUAGE, 2022, 76
  • [42] Exploring Abductive Reasoning in Language Models for Data-to-Text Generation
    Onderkova, Kristyna
    Nickles, Matthias
    2023 31ST IRISH CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COGNITIVE SCIENCE, AICS, 2023,
  • [43] Neural Pipeline for Zero-Shot Data-to-Text Generation
    Kasner, Zdenek
    Dusek, Ondrej
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 3914 - 3932
  • [44] Image and Text Fusion for Character-based Breast Cancer Classification
    Qiao, Pan
    Jin, Yanhong
    Chen, Dehua
    Zhang, YuanYuan
    IEEE 20TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING AND COMMUNICATIONS / IEEE 16TH INTERNATIONAL CONFERENCE ON SMART CITY / IEEE 4TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND SYSTEMS (HPCC/SMARTCITY/DSS), 2018, : 298 - 305
  • [45] Search and Learn: Improving Semantic Coverage for Data-to-Text Generation
    Jolly, Shailza
    Zhang, Zi Xuan
    Dengel, Andreas
    Mou, Lili
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10858 - 10866
  • [46] Navigating the text generation revolution: Traditional data-to-text NLG companies and the rise of ChatGPT
    Dale, Robert
    NATURAL LANGUAGE ENGINEERING, 2023, 29 (04) : 1188 - 1197
  • [47] DiffuD2T: Empowering Data-to-Text Generation with Diffusion
    Gong, Heng
    Feng, Xiaocheng
    Qin, Bing
    ELECTRONICS, 2023, 12 (09)
  • [48] Critic-Driven Decoding for Mitigating Hallucinations in Data-to-text Generation
    Lango, Mateusz
    Dusek, Ondrej
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 2853 - 2862
  • [49] Low-resource neural character-based noisy text normalization
    Mager, Manuel
    Jasso Rosales, Monica
    Cetinoglu, Ozlem
    Meza, Ivan
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 36 (05) : 4921 - 4929
  • [50] Template-free data-to-text generation of finnish sports news
    Kanerva, Jenna
    Rönnqvist, Samuel
    Kekki, Riina
    Salakoski, Tapio
    Ginter, Filip
    arXiv, 2019,