Ensemble-NQG-T5: Ensemble Neural Question Generation Model Based on Text-to-Text Transfer Transformer

被引:6
|
作者
Hwang, Myeong-Ha [1 ]
Shin, Jikang [1 ]
Seo, Hojin [1 ]
Im, Jeong-Seon [1 ]
Cho, Hee [1 ]
Lee, Chun-Kwon [2 ]
机构
[1] Korea Elect Power Res Inst KEPRI, Digital Solut Lab, 105 Munji Ro, Daejeon 34056, South Korea
[2] Pukyong Natl Univ, Dept Control & Instrumentat Engn, Busan 48513, South Korea
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 02期
关键词
neural question generation; deep learning; natural language processing; ensemble algorithm;
D O I
10.3390/app13020903
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Deep learning chatbot research and development is exploding recently to offer customers in numerous industries personalized services. However, human resources are used to create a learning dataset for a deep learning chatbot. In order to augment this, the idea of neural question generation (NQG) has evolved, although it has restrictions on how questions can be expressed in different ways and has a finite capacity for question generation. In this paper, we propose an ensemble-type NQG model based on the text-to-text transfer transformer (T5). Through the proposed model, the number of generated questions for each single NQG model can be greatly increased by considering the mutual similarity and the quality of the questions using the soft-voting method. For the training of the soft-voting algorithm, the evaluation score and mutual similarity score weights based on the context and the question-answer (QA) dataset are used as the threshold weight. Performance comparison results with existing T5-based NQG models using the SQuAD 2.0 dataset demonstrate the effectiveness of the proposed method for QG. The implementation of the proposed ensemble model is anticipated to span diverse industrial fields, including interactive chatbots, robotic process automation (RPA), and Internet of Things (IoT) services in the future.
引用
收藏
页数:12
相关论文
共 43 条
  • [41] SExpSMA-based T5: Serial exponential-slime mould algorithm based T5 model for question answer and distractor generation
    Bhuvan, Nikhila T.
    Jisha, G.
    Shamna, N., V
    INTELLIGENT DECISION TECHNOLOGIES-NETHERLANDS, 2024, 18 (02): : 1447 - 1462
  • [42] Neural Data-to-Text Generation Based on Small Datasets: Comparing the Added Value of Two Semi-Supervised Learning Approaches on Top of a Large Language Model
    van der Lee, Chris
    Ferreira, Thiago Castro
    Emmery, Chris
    Wiltshire, Travis J.
    Krahmer, Emiel
    COMPUTATIONAL LINGUISTICS, 2023, 49 (03) : 555 - 611
  • [43] Retraction Note: Automatic identification of drug sensitivity of cancer cell with novel regression-based ensemble convolution neural network model (Soft Computing, (2022), 26, 11, (5399-5408), 10.1007/s00500-022-07098-5)
    Gadde, Sridevi
    Charkravarthy, A.S.N.
    Satyanarayana, S.
    Murali, M.
    Soft Computing, 28 (Suppl 2):