Competitive Normalized Least-Squares Regression

被引:2
|
作者
Jamil, Waqas [1 ]
Bouchachia, Abdelhamid [1 ]
机构
[1] Bournemouth Univ, Dept Comp & Informat, Poole BH12 5BB, Dorset, England
基金
欧盟地平线“2020”;
关键词
Prediction algorithms; Approximation algorithms; Protocols; Learning systems; Covariance matrices; Time complexity; Stability analysis; Competitive analysis; least-squares; prediction; LMS ALGORITHM; EFFICIENCY; GRADIENT;
D O I
10.1109/TNNLS.2020.3009777
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Online learning has witnessed an increasing interest over the recent past due to its low computational requirements and its relevance to a broad range of streaming applications. In this brief, we focus on online regularized regression. We propose a novel efficient online regression algorithm, called online normalized least-squares (ONLS). We perform theoretical analysis by comparing the total loss of ONLS against the normalized gradient descent (NGD) algorithm and the best off-line LS predictor. We show, in particular, that ONLS allows for a better bias-variance tradeoff than those state-of-the-art gradient descent-based LS algorithms as well as a better control on the level of shrinkage of the features toward the null. Finally, we conduct an empirical study to illustrate the great performance of ONLS against some state-of-the-art algorithms using real-world data.
引用
收藏
页码:3262 / 3267
页数:6
相关论文
共 50 条