A Comprehensive Evaluation of Neural SPARQL Query Generation From Natural Language Questions

被引:0
|
作者
Diallo, Papa Abdou Karim Karou [1 ]
Reyd, Samuel [2 ]
Zouaq, Amal [1 ]
机构
[1] Polytech Montreal, Dept Comp Engn & Software Engn, LAMA WeST Lab, Montreal, PQ H3T 1J4, Canada
[2] Telecom Paris, F-91120 Palaiseau, France
来源
IEEE ACCESS | 2024年 / 12卷
基金
加拿大自然科学与工程研究理事会;
关键词
Annotations; Large language models; Computer architecture; Transformers; Vocabulary; Query processing; Knowledge based systems; Encoding; SPARQL query generation; knowledge base; copy mechanism; non pre-trained; pre-trained encoders-decoders;
D O I
10.1109/ACCESS.2024.3453215
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, the field of neural machine translation (NMT) for SPARQL query generation has witnessed significant growth. Incorporating the copy mechanism with traditional encoder-decoder architectures and using pre-trained encoder-decoder and large language models have set new performance benchmarks. This paper presents various experiments that replicate and expand upon recent NMT-based SPARQL generation studies, comparing pre-trained language models (PLMs), non-pre-trained language models (NPLMs), and large language models (LLMs), highlighting the impact of question annotation and the copy mechanism and testing various fine-tuning methods using LLMs. In particular, we provide a systematic error analysis of the models and test their generalization ability. Our study demonstrates that the copy mechanism yields significant performance enhancements for most PLMs and NPLMs. Annotating the data is pivotal to generating correct URIs, with the "tag-within" strategy emerging as the most effective approach. Additionally, our findings reveal that the primary source of errors stems from incorrect URIs in SPARQL queries that are sometimes replaced with hallucinated URIs when using base models. This does not happen using the copy mechanism, but it sometimes leads to selecting wrong URIs among candidates. Finally, the performance of the tested LLMs fell short of achieving the desired outcomes.
引用
收藏
页码:125057 / 125078
页数:22
相关论文
共 50 条
  • [21] Evaluation in Natural Language Generation: Lessons from Referring Expression Generation
    Viethen, Jette
    Dale, Robert
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2007, 48 (01): : 141 - 160
  • [22] Generation of Test Questions from RDF Files Using PYTHON']PYTHON and SPARQL
    Omarbekova, Assel
    Sharipbay, Altynbek
    Barlybaev, Alibek
    2017 INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2017), 2017, 806
  • [23] Evaluation in the context of natural language generation
    Mellish, C
    Dale, R
    COMPUTER SPEECH AND LANGUAGE, 1998, 12 (04): : 349 - 373
  • [24] Stylistic Control for Neural Natural Language Generation
    Oraby, Shereen
    COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION, 2022, : 1179 - 1179
  • [25] Enhancing Natural Language Query to SQL Query Generation Through Classification-Based Table Selection
    Chopra, Ankush
    Azam, Rauful
    ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EANN 2024, 2024, 2141 : 152 - 165
  • [26] An Overview of Natural Language Generation Systems Evaluation
    Yang, Feng-Jen
    WORLD CONGRESS ON ENGINEERING AND COMPUTER SCIENCE, WCECS 2015, VOL I, 2015, : 71 - 74
  • [27] Evaluating the Evaluation of Diversity in Natural Language Generation
    Tevet, Guy
    Berant, Jonathan
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 326 - 346
  • [28] Natural Language Generation, its Evaluation and Metrics
    Gehrmann, Sebastian
    Adewumi, Tosin
    Aggarwal, Karmanya
    Ammanamanchi, Pawan Sasanka
    Anuoluwapo, Aremu
    Bosselut, Antoine
    Chandu, Khyathi Raghavi
    Clinciu, Miruna
    Das, Dipanjan
    Dhole, Kaustubh D.
    Du, Wanyu
    Durmus, Esin
    Gangal, Varun
    Garbacea, Cristina
    Hashimoto, Tatsunori
    Hou, Yufang
    Jernite, Yacine
    Jhamtani, Harsh
    Ji, Yangfeng
    Jolly, Shailza
    Kale, Mihir
    Kumar, Dhruv
    Ladhak, Faisal
    Madaan, Aman
    Maddela, Mounica
    Mahajan, Khyati
    Mahamood, Saad
    Majumder, Bodhisattwa Prasad
    Martins, Pedro Henrique
    McMillan-Major, Angelina
    Mille, Simon
    van Miltenburg, Emiel
    Nadeem, Moin
    Narayan, Shashi
    Nikolaev, Vitaly
    Niyongabo, Rubungo Andre
    Osei, Salomey
    Parikh, Ankur
    Perez-Beltrachini, Laura
    Rao, Niranjan Ramesh
    Raunak, Vikas
    Rodriguez, Juan Diego
    Santhanam, Sashank
    Sedoc, Joao
    Sellam, Thibault
    Shaikh, Samira
    Shimorina, Anastasia
    Sobrevilla Cabezudo, Marco Antonio
    Strobelt, Hendrik
    Subramani, Nishant
    1ST WORKSHOP ON NATURAL LANGUAGE GENERATION, EVALUATION, AND METRICS (GEM 2021), 2021, : 96 - 120
  • [29] Evaluation of African American Language Bias in Natural Language Generation
    Deas, Nicholas
    Grieser, Jessi
    Kleiner, Shana
    Patton, Desmond
    Turcan, Elsbeth
    McKeown, Kathleen
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 6805 - 6824
  • [30] Data Agnostic RoBERTa-based Natural Language to SQL Query Generation
    Pal, Debaditya
    Sharma, Harsh
    Chaudhuri, Kaustubh
    2021 6TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2021,