A statistical learning theory approach of bloat

被引:0
|
作者
Gelly, Sylvain [1 ]
Teytaud, Olivier [1 ]
Bredeche, Nicolas [1 ]
Schoenauer, Marc [1 ]
机构
[1] Univ Paris 11, INRIA Futurs, Equipe TAO, LRI, F-91405 Orsay, France
关键词
algorithms; reliability; theory; code bloat; code growth; Genetic Programming; Statistical Learning Theory;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Code bloat, the excessive increase of code size, is an important issue in Genetic Programming (GP). This paper proposes a theoretical analysis of code bloat in the framework of symbolic regression in GP, from the viewpoint of Statistical Learning Theory, a well grounded mathematical toolbox for Machine Learning. Two kinds of bloat must be distinguished in that context, depending whether the target function lies in the search space or not. Then, important mathematical results are proved using classical results from Statistical Learning. Namely, the Vapnik-Chervonenkis dimension of programs is computed, and further results from Statistical Learning allow to prove that a parsimonious fitness ensures Universal Consistency (the solution minimizing the empirical error does converge to the best possible error when the number of examples goes to infinity). However, it is proved that the standard method consisting in choosing a maximal program size depending on the number of examples might still result in programs of infinitely increasing size with their accuracy; a more complicated modification of the fitness is proposed that theoretically avoids unnecessary bloat while nevertheless preserving the Universal Consistency.
引用
收藏
页码:1783 / 1784
页数:2
相关论文
共 50 条