Improving hyper-parameter self-tuning for data streams by adapting an evolutionary approach

被引:1
|
作者
Moya, Antonio R. [1 ]
Veloso, Bruno [2 ,3 ]
Gama, Joao [2 ,3 ]
Ventura, Sebastian [1 ]
机构
[1] Univ Cordoba, Andalusian Res Inst Data Sci & Computat Intelligen, Dept Comp Sci & Numer Anal, Cordoba, Spain
[2] INESC TEC, Porto, Portugal
[3] Univ Porto, FEP, Porto, Portugal
关键词
Data streams; Concept drift; Optimisation; Hyper-parameters; Evolutionary algorithms; ALGORITHMS;
D O I
10.1007/s10618-023-00997-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hyper-parameter tuning of machine learning models has become a crucial task in achieving optimal results in terms of performance. Several researchers have explored the optimisation task during the last decades to reach a state-of-the-art method. However, most of them focus on batch or offline learning, where data distributions do not change arbitrarily over time. On the other hand, dealing with data streams and online learning is a challenging problem. In fact, the higher the technology goes, the greater the importance of sophisticated techniques to process these data streams. Thus, improving hyper-parameter self-tuning during online learning of these machine learning models is crucial. To this end, in this paper, we present MESSPT, an evolutionary algorithm for self-hyper-parameter tuning for data streams. We apply Differential Evolution to dynamically-sized samples, requiring a single pass-over of data to train and evaluate models and choose the best configurations. We take care of the number of configurations to be evaluated, which necessarily has to be reduced, thus making this evolutionary approach a micro-evolutionary one. Furthermore, we control how our evolutionary algorithm deals with concept drift. Experiments on different learning tasks and over well-known datasets show that our proposed MESSPT outperforms the state-of-the-art on hyper-parameter tuning for data streams.
引用
收藏
页码:1289 / 1315
页数:27
相关论文
共 50 条
  • [21] A hyper-parameter tuning approach for cost-sensitive support vector machine classifiers
    Guido, Rosita
    Groccia, Maria Carmela
    Conforti, Domenico
    SOFT COMPUTING, 2023, 27 (18) : 12863 - 12881
  • [22] A Modified Bayesian Optimization based Hyper-Parameter Tuning Approach for Extreme Gradient Boosting
    Putatunda, Sayan
    Rama, Kiran
    2019 FIFTEENTH INTERNATIONAL CONFERENCE ON INFORMATION PROCESSING (ICINPRO): INTERNET OF THINGS, 2019, : 6 - 11
  • [23] A pragmatic approach for hyper-parameter tuning in search-based test case generation
    Shayan Zamani
    Hadi Hemmati
    Empirical Software Engineering, 2021, 26
  • [24] A hyper-parameter tuning approach for cost-sensitive support vector machine classifiers
    Rosita Guido
    Maria Carmela Groccia
    Domenico Conforti
    Soft Computing, 2023, 27 : 12863 - 12881
  • [25] Hyper-Parameter Optimization for Improving the Performance of Grammatical Evolution
    Wang, Hao
    Lou, Yitan
    Back, Thomas
    2019 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2019, : 2649 - 2656
  • [26] Airline Price Prediction Using XGBoost Hyper-parameter Tuning
    Kumar, Amit
    ADVANCED NETWORK TECHNOLOGIES AND INTELLIGENT COMPUTING, ANTIC 2022, PT II, 2023, 1798 : 239 - 248
  • [27] Self-Tuning, Bandwidth-Aware Monitoring for Dynamic Data Streams
    Jain, Navendu
    Yalagandula, Praveen
    Dahlin, Mike
    Zhang, Yin
    ICDE: 2009 IEEE 25TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING, VOLS 1-3, 2009, : 114 - 125
  • [28] A Self-Tuning Fuzzy Rule-Based Classifier for Data Streams
    Shahparast, Homeira
    Hamzeloo, Sam
    Jahromi, Mansoor Zolghadri
    INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2014, 22 (02) : 293 - 303
  • [29] Self-Adapting Approach in Parameter Tuning for Differential Evolution
    Wang, Shir Li
    Theam Foo Ng
    Jamil, Nurul Aini
    Samuri, Suzani Mohamad
    Mailok, Ramlah
    Rahmatullah, Bahbibi
    2015 CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE (TAAI), 2015, : 113 - 119
  • [30] AUTOMATA : Gradient Based Data Subset Selection for Compute-Efficient Hyper-parameter Tuning
    Killamsetty, Krishnateja
    Abhishek, Guttu Sai
    Aakriti
    Ramakrishnan, Ganesh
    Evfimievski, Alexandre V.
    Popa, Lucian
    Iyer, Rishabh
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,