Likelihood-based cross-validation score for selecting the smoothing parameter in maximum penalized likelihood estimation

被引:0
|
作者
Sakamoto, W [1 ]
Shirahata, S [1 ]
机构
[1] Osaka Univ, Grad Sch Engn Sci, Dept Informat & Math Sci, Toyonaka, Osaka 5608531, Japan
关键词
Akaike information criterion; logistic regression; nonparametric generalized linear models; Poisson regression; smoothing spline;
D O I
10.1080/03610929908832379
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Maximum penalized likelihood estimation is applied in non(semi)-parametric regression problems, and enables us exploratory identification and diagnostics of nonlinear regression relationships. The smoothing parameter lambda controls trade-off between the smoothness and the goodness-of-fit oof a function. The method of cross-validation is used for selecting lambda, but the generalized cross-validation, which is based on the squared error criterion, shows bad behavior in non-normal distribution and can not often select reasonable lambda. The purpose of this study is to propose a method which gives more suitable lambda and to evaluate the performance of it. A method of simple calculation for the delete-one estimates in the likelihood-based cross-validation (LCV) score is described. A score of similar form to the Akaike information criterion (AIC) is also derived. The proposed scores are compared with the ones of standard procedures by using data sets in literatures. Simulations are performed to compare the patterns of selecting lambda and overall goodness-of-fit and to evaluate the effects of some factors. The LCV scares by the simple calculation provide good approximations to the exact one if lambda is not extremely small. Furthermore the LCV scores by the simple calculation have little risk of choosing extremely small lambda and make it possible to select lambda adaptively. They have the effect of reducing the bias of estimates and provide better performance in the sense of overall goodness of-fit. These scores are useful especially in the case of small sample size and in the case of binary logistic regression.
引用
收藏
页码:1671 / 1698
页数:28
相关论文
共 50 条