Learning Large Scale Ordinal Ranking Model via Divide-and-Conquer Technique

被引:0
|
作者
Tang, Lu [1 ]
Chaudhuri, Sougata [2 ]
Bagherjeiran, Abraham [2 ]
Zhou, Ling [1 ]
机构
[1] Univ Michigan, Ann Arbor, MI 48109 USA
[2] A9 Com Inc, Palo Alto, CA USA
关键词
Binary Classification; Ordinal Ranking; Big Data;
D O I
10.1145/3184558.3191658
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Structured prediction, where outcomes have a precedence order, lies at the heart of machine learning for information retrieval, movie recommendation, product review prediction, and digital advertising. Ordinal ranking, in particular, assumes that the structured response has a linear ranked order. Due to the extensive applicability of these models, substantial research has been devoted to understanding them, as well as developing efficient training techniques. One popular and widely cited technique of training ordinal ranking models is to exploit the linear precedence order and systematically reduce it to a binary classification problem. This facilitates the usage of readily available, powerful binary classifiers, but necessitates an expansion of the original training data, where the training data increases by K - 1 times of its original size, with K being the number of ordinal classes. Due to prevalent nature of problems with large number of ordered classes, the reduction leads to datasets which are too large to train on single machines. While approximation methods like stochastic gradient descent are typically applied here, we investigate exact optimization solutions that can scale. In this paper, we present a divide-and-conquer (DC) algorithm, which divides large scale binary classification data into a cluster of machines and trains logistic models in parallel, and combines them at the end of the training phase to create a single binary classifier, which can then be used as an ordinal ranker. It requires no synchronization between the parallel learning algorithms during the training period, which makes training on large datasets feasible and efficient. We prove consistency and asymptotic normality property of the learned models using our proposed algorithm. We provide empirical evidence, on various ordinal datasets, of improved estimation and prediction performance of the model learnt using our algorithm, over several standard divide-and-conquer algorithms.
引用
收藏
页码:1901 / 1909
页数:9
相关论文
共 50 条
  • [41] Divide-and-conquer large scale capacitated arc routing problems with route cutting off decomposition
    Zhang, Yuzhou
    Mei, Yi
    Zhang, Buzhong
    Jiang, Keqin
    INFORMATION SCIENCES, 2021, 553 : 208 - 224
  • [42] Model selection via Bayesian information criterion for divide-and-conquer penalized quantile regression
    Kang, Jongkyeong
    Han, Seokwon
    Bang, Sungwan
    KOREAN JOURNAL OF APPLIED STATISTICS, 2022, 35 (02) : 217 - 227
  • [43] Random sampling of contingency tables via probabilistic divide-and-conquer
    DeSalvo, Stephen
    Zhao, James
    COMPUTATIONAL STATISTICS, 2020, 35 (02) : 837 - 869
  • [44] A divide-and-conquer machine learning approach for modeling turbulent flows
    Man, Anthony
    Jadidi, Mohammad
    Keshmiri, Amir
    Yin, Hujun
    Mahmoudi, Yasser
    PHYSICS OF FLUIDS, 2023, 35 (05)
  • [45] A Divide-and-Conquer Learning Approach to Radial Basis Function Networks
    YIU-MING CHEUNG
    RONG-BO HUANG
    Neural Processing Letters, 2005, 21 : 189 - 206
  • [46] A DIVIDE-AND-CONQUER DICTIONARY LEARNING ALGORITHM AND ITS PERFORMANCE ANALYSIS
    Mukherjee, Subhadip
    Seelamantula, Chandra Sekhar
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 4712 - 4716
  • [47] Few-Shot Segmentation via Divide-and-Conquer Proxies
    Chunbo Lang
    Gong Cheng
    Binfei Tu
    Junwei Han
    International Journal of Computer Vision, 2024, 132 : 261 - 283
  • [48] A divide-and-conquer learning approach to radial basis function networks
    Cheung, YM
    Huang, RB
    NEURAL PROCESSING LETTERS, 2005, 21 (03) : 189 - 206
  • [49] FPGA Divide-and-Conquer Placement using Deep Reinforcement Learning
    Wang, Shang
    Mamillapalli, Deepak Ranganatha Sastry
    Yang, Tianpei
    Taylor, Matthew E.
    2024 INTERNATIONAL SYMPOSIUM OF ELECTRONICS DESIGN AUTOMATION, ISEDA 2024, 2024, : 690 - 696
  • [50] Automated learning of RVM for large scale text sets: Divide to conquer
    Silva, Catarina
    Ribeiro, Bernardete
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2006, PROCEEDINGS, 2006, 4224 : 878 - 886