Few-Shot Tabular Data Enrichment Using Fine-Tuned Transformer Architectures

被引:0
|
作者
Harari, Asaf [1 ]
Katz, Gilad [1 ]
机构
[1] Ben Gurion Univ Negev, POB 653, Beer Sheva, Israel
关键词
KNOWLEDGE-BASE; DBPEDIA;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The enrichment of tabular datasets using external sources has gained significant attention in recent years. Existing solutions, however, either ignore external unstructured data completely or devise dataset-specific solutions. In this study, we proposed Few-Shot Transformer based Enrichment (FeSTE), a generic and robust framework for the enrichment of tabular datasets using unstructured data. By training over multiple datasets, our approach is able to develop generic models that can be applied to additional datasets with minimal training (i.e., few-shot). Our approach is based on an adaptation of BERT, for which we present a novel fine-tuning approach that reformulates the tuples of the datasets as sentences. Our evaluation, conducted on 17 datasets, shows that FeSTE is able to generate high quality features and significantly outperform existing fine-tuning solutions.
引用
收藏
页码:1577 / 1591
页数:15
相关论文
共 50 条
  • [1] On the Utility of Domain-Adjacent Fine-Tuned Model Ensembles for Few-shot Problems
    Alam, Md Ibrahim Ibne
    Kar, Koushik
    Ram, Parikshit
    Dan, Soham
    Samulowitz, Horst
    arXiv,
  • [2] Disaster Tweet Classification Using Fine-Tuned Deep Learning Models Versus Zero and Few-Shot Large Language Models
    Dinani, Soudabeh Taghian
    Caragea, Doina
    Gyawali, Nikesh
    DATA MANAGEMENT TECHNOLOGIES AND APPLICATIONS, DATA 2023, 2024, 2105 : 73 - 94
  • [3] Few-Shot Steel Defect Detection Based on a Fine-Tuned Network with Serial Multi-Scale Attention
    Liu, Xiangpeng
    Jiao, Lei
    Peng, Yulin
    An, Kang
    Wang, Danning
    Lu, Wei
    Han, Jianjiao
    APPLIED SCIENCES-BASEL, 2024, 14 (13):
  • [4] Structuring medication signeturs as a language regression task: comparison of zero- and few-shot GPT with fine-tuned models
    Garcia-Agundez, Augusto
    Kay, Julia L.
    Li, Jing
    Gianfrancesco, Milena
    Rai, Baljeet
    Hu, Angela
    Schmajuk, Gabriela
    Yazdany, Jinoos
    JAMIA OPEN, 2024, 7 (02)
  • [5] TabLLM: Few-shot Classification of Tabular Data with Large Language Models
    Hegselmann, Stefan
    Buendia, Alejandro
    Lang, Hunter
    Agrawal, Monica
    Jiang, Xiaoyi
    Sontag, David
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [6] Fine-tuned LLMs Know More, Hallucinate Less with Few-Shot Sequence-to-Sequence Semantic Parsing over Wikidata
    Xu, Silei
    Liu, Shicheng
    Culhane, Theo
    Pertseva, Elizaveta
    Wu, Meng-Hsi
    Semnani, Sina J.
    Lam, Monica S.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 5778 - 5791
  • [7] Fine-Tuned Transformer Model for Sentiment Analysis
    Liu, Sishun
    Shuai, Pengju
    Zhang, Xiaowu
    Chen, Shuang
    Li, Li
    Liu, Ming
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT (KSEM 2020), PT II, 2020, 12275 : 336 - 343
  • [8] Adaptive Agent Transformer for Few-Shot Segmentation
    Wang, Yuan
    Sun, Rui
    Zhang, Zhe
    Zhang, Tianzhu
    COMPUTER VISION, ECCV 2022, PT XXIX, 2022, 13689 : 36 - 52
  • [9] Dynamic Transformer for Few-shot Instance Segmentation
    Wang, Haochen
    Liu, Jie
    Liu, Yongtuo
    Maji, Subhransu
    Sonke, Jan-Jakob
    Gavves, Efstratios
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 2969 - 2977
  • [10] Mask Matching Transformer for Few-Shot Segmentation
    Jiao, Siyu
    Zhang, Gengwei
    Navasardyan, Shant
    Chen, Ling
    Zhao, Yao
    Wei, Yunchao
    Shi, Humphrey
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,