Regret Bounds for Online Pairwise Learning with Non-Convex Loss Functions Using Stability Analysis

被引:0
|
作者
Lang X. [1 ]
Li C. [1 ,2 ]
Liu Y. [3 ,4 ]
Wang M. [1 ,2 ]
机构
[1] College of Computer and Information Technology, Northeastern Petroleum University, Heilongjiang, Daqing
[2] Heilongjiang Provincial Key Laboratory of Petroleum Big Data and Intelligent Analysis, Northeastern Petroleum University, Heilongjiang, Daqing
[3] Gaoling School of Artificial Intelligence, Renmin University of China, Beijing
[4] Beijing Key Laboratory of Big Data Management and Analysis Methods, Renmin University of China, Beijing
基金
中国国家自然科学基金;
关键词
non-convex; offline optimization oracle; online pairwise learning; regret bounds; stability;
D O I
10.7544/issn1000-1239.202220221
中图分类号
学科分类号
摘要
Pairwise learning refers to a learning task which involves a loss function depending on pairs of instances. Recently, there is a growing interest in studying pairwise learning since it includes many important machine learning tasks as specific examples, e.g., metric learning, AUC maximization and ranking. Regret bounds are particularly important for generalization analysis of online pairwise learning. The existing online pairwise learning analysis provides regret bounds only with convex loss functions. To fill the gap in the theoretical study of online pairwise learning with non-convex loss functions, we present a systematic study on the generalization analysis for online pairwise learning and propose regret bounds for non-convex online pairwise learning in this paper. We consider online learning in an adversarial, non-convex setting under the assumption that the learner has access to an offline optimization oracle and the learner’s prediction with expert advice. We first propose a general online pairwise learning framework and establish the stability of online pairwise learning with non-convex loss functions. Then, the regret bounds can be derived naturally from stability. Finally, we show that the general online pairwise learning framework with non-convex loss functions achieves optimal regret bounds of O(T−1/2) when the learner has access to an offline optimization oracle. © 2023 Science Press. All rights reserved.
引用
收藏
页码:2806 / 2813
页数:7
相关论文
共 26 条
  • [1] Zhijie Li, Yuanxiang Li, Feng Wang, Et al., Online learning algorithms for big data analytics: A survey[J], Journal of Computer Research and Development, 52, 8, (2015)
  • [2] Clemencon S, Lugosi G, Vayatis N., Ranking and empirical minimization of U-statistics[J], The Annals of Statistics, 36, 2, pp. 844-874, (2008)
  • [3] Agarwal S, Niyogi P., Generalization bounds for ranking algorithms via algorithmic stability[J], Journal of Machine Learning Research, 10, 2, pp. 441-474, (2009)
  • [4] Rejchel W., On ranking and generalization bounds[J], Journal of Machine Learning Research, 13, 5, pp. 1373-1392, (2012)
  • [5] Rejchel W., Fast rates for ranking with large families[J], Neurocomputing, 168, pp. 1104-1110, (2015)
  • [6] Peilin Zhao, Hoi S, Jin Rong, Et al., Online AUC maximization[C], Proc of the 28th Int Conf on Machine Learning, pp. 233-240, (2011)
  • [7] Weinberger K Q, Saul L K., Distance metric learning for large margin nearest neighbor classification[J], Journal of Machine Learning Research, 10, 2, pp. 207-244, (2009)
  • [8] Yuyang Wang, Khardon R, Pechyony D, Et al., Generalization bounds for online learning algorithms with pairwise loss functions[C], Proc of the 25th Conf on Learning Theory, pp. 1-22, (2012)
  • [9] Yiming Ying, Dingxuan Zhou, Unregularized online learning algorithms with general loss functions[J], Applied and Computational Harmonic Analysis, 42, 2, pp. 224-244, (2017)
  • [10] Yiming Ying, Dingxuan Zhou, Online pairwise learning algorithms[J], Neural Computation, 28, 4, pp. 743-777, (2016)