Scaling Up Differentially Private LASSO Regularized Logistic Regression via Faster Frank-Wolfe Iterations

被引:0
|
作者
Raff, Edward [1 ]
Khanna, Amol [1 ]
Lu, Fred [1 ]
机构
[1] Booz Allen Hamilton, Mclean, VA 22102 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To the best of our knowledge, there are no methods today for training differentially private regression models on sparse input data. To remedy this, we adapt the Frank-Wolfe algorithm for L-1 penalized linear regression to be aware of sparse inputs and to use them effectively. In doing so, we reduce the training time of the algorithm from O(TDS + TNS) to O(NS + T root D log D + TS2), where T is the number of iterations and a sparsity rate S of a dataset with N rows and D features. Our results demonstrate that this procedure can reduce runtime by a factor of up to 2, 200x, depending on the value of the privacy parameter epsilon and the sparsity of the dataset.
引用
收藏
页数:15
相关论文
共 1 条
  • [1] Fast and scalable Lasso via stochastic Frank-Wolfe methods with a convergence guarantee
    Frandi, Emanuele
    Nanculef, Ricardo
    Lodi, Stefano
    Sartori, Claudio
    Suykens, Johan A. K.
    MACHINE LEARNING, 2016, 104 (2-3) : 195 - 221