Leveraging Large Language Models for Flexible and Robust Table-to-Text Generation

被引:0
|
作者
Oro, Ermelinda [1 ,2 ]
De Grandis, Luca [2 ]
Granata, Francesco Maria [2 ]
Ruffolo, Massimo [1 ,2 ]
机构
[1] CNR, Inst High Performance Comp & Networking, Via P Bucci 8-9C, I-87036 Arcavacata Di Rende, CS, Italy
[2] Univ Calabria, TechNest Start Incubator, Altilia Srl, Piazza Vermicelli, I-87036 Arcavacata Di Rende, CS, Italy
关键词
Natural Language Generation; Table-to-Text; Data-to-Text; LLM; Zero-Shot; GPT-3; LLaMa; Prompt; Finetuning; LoRA;
D O I
10.1007/978-3-031-68309-1_19
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Generating natural language descriptions from structured tabular data is a crucial challenge with high-impact applications across diverse domains, including business intelligence, scientific communication, and data analytics. Traditional rule-based and machine learning approaches have faced limitations in reusability, vocabulary coverage, and handling complex table layouts. Recent advances in LLMs pre-trained on vast corpora offer an opportunity to overcome these limitations by leveraging their strong language understanding and generation capabilities in a flexible learning setup. In this paper, We conduct a comprehensive evaluation of two LLMs - GPT-3.5 and LLaMa2-7B - on table-to-text generation across three diverse public datasets: WebNLG, NumericNLG, and ToTTo. Our experiments investigate both zero-shot prompting techniques and finetuning using the parameter-efficient LoRA method. Results demonstrate GPT-3.5's impressive capabilities, outperforming LLaMa2 in zero-shot settings. However, finetuning LLaMa2 on a subset of data significantly bridges this performance gap and produces generations much closer to ground truth and comparable to SOTA approaches. Our findings highlight LLMs' promising potential for data-to-text while identifying key areas for future research.
引用
收藏
页码:222 / 227
页数:6
相关论文
共 50 条
  • [1] Enabling controllable table-to-text generation via prompting large language models with guided planning
    Zhao, Shuo
    Sun, Xin
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [2] Table-to-Text Generation With Pretrained Diffusion Models
    Krylov, Aleksei S.
    Somov, Oleg D.
    IEEE ACCESS, 2024, 12 : 110517 - 110525
  • [3] Table-to-Text: Describing Table Region with Natural Language
    Bao, Junwei
    Tang, Duyu
    Duan, Nan
    Yan, Zhao
    Lv, Yuanhua
    Zhou, Ming
    Zhao, Tiejun
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 5020 - 5027
  • [4] Towards Table-to-Text Generation with Numerical Reasoning
    Suadaa, Lya Hulliyyatus
    Kamigaito, Hidetaka
    Funakoshi, Kotaro
    Okumura, Manabu
    Takamura, Hiroya
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 1451 - 1465
  • [5] Robust (Controlled) Table-to-Text Generation with Structure-Aware Equivariance Learning
    Wang, Fei
    Xu, Zhewei
    Szekely, Pedro
    Chen, Muhao
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5037 - 5048
  • [6] Generative adversarial network for Table-to-Text generation
    Zhao, Jianyu
    Zhan, Zhiqiang
    Li, Tong
    Li, Rang
    Hu, Changjian
    Wang, Siyun
    Zhang, Yang
    NEUROCOMPUTING, 2021, 452 : 28 - 36
  • [7] ToTTo: A Controlled Table-To-Text Generation Dataset
    Parikh, Ankur P.
    Wang, Xuezhi
    Gehrmann, Sebastian
    Faruqui, Manaal
    Dhingra, Bhuwan
    Yang, Diyi
    Das, Dipanjan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1173 - 1186
  • [8] TableSF: A Structural Bias Framework for Table-To-Text Generation
    Liu, Di
    Wang, Weihua
    Bao, Feilong
    Gao, Guanglai
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IX, 2023, 14262 : 401 - 412
  • [9] Learning number reasoning for numerical table-to-text generation
    Feng, Xiaocheng
    Gong, Heng
    Chen, Yuyu
    Sun, Yawei
    Qin, Bing
    Bi, Wei
    Liu, Xiaojiang
    Liu, Ting
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (08) : 2269 - 2280
  • [10] Learning number reasoning for numerical table-to-text generation
    Feng, Xiaocheng
    Gong, Heng
    Chen, Yuyu
    Sun, Yawei
    Qin, Bing
    Bi, Wei
    Liu, Xiaojiang
    Liu, Ting
    International Journal of Machine Learning and Cybernetics, 2021, 12 (08): : 2269 - 2280