Parallel Large-Scale Neural Network Training For Online Advertising

被引:0
|
作者
Qi, Quanchang [1 ]
Lu, Guangming [1 ]
Zhang, Jun [1 ]
Yang, Lichun [1 ]
Liu, Haishan [1 ]
机构
[1] Tencent Ads, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks have shown great successes in many fields. Due to the complexity of the training pipeline, however, using them in an industrial setting is challenging. In online advertising, the complexity arises from the immense size of the training data, and the dimensionality of the sparse feature space (both can be hundreds of billions). To tackle these challenges, we built TrainSparse (TS), a system that parallelizes the training of neural networks with a focus on efficiently handling large-scale sparse features. In this paper, we present the design and implementation of TS, and show the effectiveness of the system by applying it to predict the ad conversion rate (pCVR), one of the key problems in online advertising. We also compare several methods for dimensionality reduction on sparse features in the pCVR task. Experiments on real-world industry data show that TS achieves outstanding performance and scalability.
引用
收藏
页码:343 / 350
页数:8
相关论文
共 50 条
  • [31] Efficient on-chip training of large-scale optical neural network through block adjoint training algorithm
    Yang, Zhiwei
    Zhang, Tian
    Dai, Jian
    Xu, Kun
    OPTICS EXPRESS, 2024, 32 (26): : 46633 - 46648
  • [32] Marginalized Neural Network Mixtures for Large-Scale Regression
    Lazaro-Gredilla, Miguel
    Figueiras-Vidal, Anibal R.
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2010, 21 (08): : 1345 - 1351
  • [33] ESTIMATING ADVANTAGES TO LARGE-SCALE ADVERTISING
    BROWN, RS
    REVIEW OF ECONOMICS AND STATISTICS, 1978, 60 (03) : 428 - 437
  • [34] Understanding the Implication of Non-Volatile Memory for Large-Scale Graph Neural Network Training
    Lee, Yunjae
    Kwon, Youngeun
    Rhu, Minsoo
    IEEE COMPUTER ARCHITECTURE LETTERS, 2021, 20 (02) : 118 - 121
  • [35] THE DUAL BACKBONE NETWORK - DISTRIBUTED AND PARALLEL PROCESSING ON A LARGE-SCALE
    ENDRIZZI, A
    COMPUTER NETWORKS AND ISDN SYSTEMS, 1987, 14 (2-5): : 373 - 381
  • [36] Enabling Parallel Simulation of Large-Scale HPC Network Systems
    Mubarak, Misbah
    Carothers, Christopher D.
    Ross, Robert B.
    Carns, Philip
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2017, 28 (01) : 87 - 100
  • [37] paraSNF: An Parallel Approach for Large-Scale Similarity Network Fusion
    Shen, Xiaolong
    He, Song
    Fang, Minquan
    Wen, Yuqi
    Bo, Xiaochen
    Dou, Yong
    ADVANCED COMPUTER ARCHITECTURE, 2018, 908 : 155 - 167
  • [38] Robust Parallel Pursuit for Large-Scale Association Network Learning
    Li, Wenhui
    Zhou, Xin
    Dong, Ruipeng
    Zheng, Zemin
    INFORMS JOURNAL ON COMPUTING, 2024,
  • [39] Online training of parallel neural network estimators for control of induction motors
    Rubaai, A
    Kotaru, R
    Kankam, MD
    IEEE TRANSACTIONS ON INDUSTRY APPLICATIONS, 2001, 37 (05) : 1512 - 1521
  • [40] Training of large-scale feed-forward neural networks
    Seiffert, Udo
    2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10, 2006, : 5324 - 5329