Balancing Efficiency and Performance in NLP: A Cross-Comparison of Shallow Machine Learning and Large Language Models via AutoML

被引:0
|
作者
Estevanell-Valladares, Ernesto L. [1 ,2 ]
Gutierrez, Yoan [2 ]
Montoyo-Guijarro, Andres [2 ]
Munoz-Guillena, Rafael [2 ]
Almeida-Cruz, Yudivian [1 ]
机构
[1] Univ Habana, Havana, Cuba
[2] Univ Alicante, Alicante, Spain
来源
关键词
Natural Language Processing; Machine Learning; AutoML; LLM;
D O I
10.26342/2024-73-16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study critically examines the resource efficiency and performance ofShallow Machine Learning (SML) methods versus Large Language Models (LLMs)in text classification tasks by exploring the balance between accuracy and environ-mental sustainability. We introduce a novel optimization strategy that prioritizescomputational efficiency and ecological impact alongside traditional performancemetrics leveraging Automated Machine Learning (AutoML). Our analysis revealsthat while the pipelines we developed did not surpass state-of-the-art (SOTA) modelsregarding raw performance, they offer a significantly reduced carbon footprint. Wediscovered SML optimal pipelines with competitive performance and up to 70 timesless carbon emissions than hybrid or fully LLM pipelines, such as standard BERT andDistilBERT variants. Similarly, we obtain hybrid pipelines (using SML and LLMs)with between 20% and 50% reduced carbon emissions compared to fine-tuned alter-natives and only a marginal decrease in performance. This research challenges theprevailing reliance on computationally intensive LLMs for NLP tasks and underscoresthe untapped potential of AutoML in sculpting the next wave of environmentallyconscious AI models.
引用
收藏
页码:221 / 233
页数:13
相关论文
共 6 条
  • [1] Comparison of Machine Learning Algorithms and Large Language Models for Product Categorization
    Ihsanoglu, Abdullah
    Zaval, Mounes
    Yildiz, Olcay Taner
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [2] Comparison of Large Language Models in Generating Machine Learning Curricula in High Schools
    Noveski, Gjorgji
    Jeroncic, Mathis
    Velard, Thomas
    Kocuvan, Primoz
    Gams, Matjaz
    ELECTRONICS, 2024, 13 (20)
  • [3] Comparison between Machine-Learning-Based Turbidity Models Developed for Different Lake Zones in a Large Shallow Lake
    Hu, Runtao
    Xu, Wangchen
    Yan, Wenming
    Wu, Tingfeng
    He, Xiangyu
    Cheng, Nannan
    WATER, 2023, 15 (03)
  • [4] Context-Aware Behavioral Tips to Improve Sleep Quality via Machine Learning and Large Language Models
    Corda, Erica
    Massa, Silvia M.
    Riboni, Daniele
    FUTURE INTERNET, 2024, 16 (02)
  • [5] Comparing human text classification performance and explainability with large language and machine learning models using eye-tracking
    Venkatesh, Jeevithashree Divya
    Jaiswal, Aparajita
    Nanda, Gaurav
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [6] Forecasting performance comparison of two hybrid machine learning models for cooling load of a large-scale commercial building
    Zhou Xuan
    Zi Xuehui
    Liang Liequan
    Fan Zubing
    Yan Junwei
    Pan Dongmei
    JOURNAL OF BUILDING ENGINEERING, 2019, 21 : 64 - 73