Learning Large Scale Ordinal Ranking Model via Divide-and-Conquer Technique

被引:0
|
作者
Tang, Lu [1 ]
Chaudhuri, Sougata [2 ]
Bagherjeiran, Abraham [2 ]
Zhou, Ling [1 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
[2] A9 Com Inc, Palo Alto, CA USA
关键词
Binary Classification; Ordinal Ranking; Big Data;
D O I
10.1145/3184558.3191658
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Structured prediction, where outcomes have a precedence order, lies at the heart of machine learning for information retrieval, movie recommendation, product review prediction, and digital advertising. Ordinal ranking, in particular, assumes that the structured response has a linear ranked order. Due to the extensive applicability of these models, substantial research has been devoted to understanding them, as well as developing efficient training techniques. One popular and widely cited technique of training ordinal ranking models is to exploit the linear precedence order and systematically reduce it to a binary classification problem. This facilitates the usage of readily available, powerful binary classifiers, but necessitates an expansion of the original training data, where the training data increases by K - 1 times of its original size, with K being the number of ordinal classes. Due to prevalent nature of problems with large number of ordered classes, the reduction leads to datasets which are too large to train on single machines. While approximation methods like stochastic gradient descent are typically applied here, we investigate exact optimization solutions that can scale. In this paper, we present a divide-and-conquer (DC) algorithm, which divides large scale binary classification data into a cluster of machines and trains logistic models in parallel, and combines them at the end of the training phase to create a single binary classifier, which can then be used as an ordinal ranker. It requires no synchronization between the parallel learning algorithms during the training period, which makes training on large datasets feasible and efficient. We prove consistency and asymptotic normality property of the learned models using our proposed algorithm. We provide empirical evidence, on various ordinal datasets, of improved estimation and prediction performance of the model learnt using our algorithm, over several standard divide-and-conquer algorithms.
引用
收藏
页码:1901 / 1909
页数:9
相关论文
共 50 条
  • [1] Iteratively Divide-and-Conquer Learning for Nonlinear Classification and Ranking
    Wu, Ou
    Mao, Xue
    Hu, Weiming
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2018, 9 (02)
  • [2] Large-Scale Quantum Approximate Optimization via Divide-and-Conquer
    Li, Junde
    Alam, Mahabubul
    Ghosh, Swaroop
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (06) : 1852 - 1860
  • [3] A Divide-and-Conquer Approach for Large-scale Multi-label Learning
    Zhang, Wenjie
    Wang, Xiangfeng
    Yan, Junchi
    Zha, Hongyuan
    2017 IEEE THIRD INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2017), 2017, : 398 - 401
  • [4] DIVIDE-AND-CONQUER TOMOGRAPHY FOR LARGE-SCALE NETWORKS
    Santos, Augusto
    Matta, Vincenzo
    Sayed, Ali H.
    2018 IEEE DATA SCIENCE WORKSHOP (DSW), 2018, : 170 - 174
  • [5] Divide-and-conquer based large-scale spectral clustering
    Li, Hongmin
    Ye, Xiucai
    Imakura, Akira
    Sakurai, Tetsuya
    NEUROCOMPUTING, 2022, 501 : 664 - 678
  • [6] A Divide-and-Conquer Framework for Large-Scale Subspace Clustering
    You, Chong
    Donnat, Claire
    Robinson, Daniel P.
    Vidal, Rene
    2016 50TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS AND COMPUTERS, 2016, : 1014 - 1018
  • [7] An eigenspace divide-and-conquer approach for large-scale optimization
    Ren, Zhigang
    Liang, Yongsheng
    Wang, Muyi
    Yang, Yang
    Chen, An
    APPLIED SOFT COMPUTING, 2021, 99
  • [8] Divide-and-Conquer Strategies for Large-scale Simulations in R
    Zhang, Hui
    Zhong, Yiwen
    Lin, Juan
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 3517 - 3523
  • [9] A divide-and-conquer method for large scale ν-nonparallel support vector machines
    Xuchan Ju
    Yingjie Tian
    Neural Computing and Applications, 2018, 29 : 497 - 509
  • [10] A divide-and-conquer method for large scale ν-nonparallel support vector machines
    Ju, Xuchan
    Tian, Yingjie
    NEURAL COMPUTING & APPLICATIONS, 2018, 29 (09): : 497 - 509