A MODEL OF DOUBLE DESCENT FOR HIGH-DIMENSIONAL LOGISTIC REGRESSION

被引:0
|
作者
Deng, Zeyu [1 ]
Kammoun, Abla [2 ]
Thrampoulidis, Christos [1 ]
机构
[1] Univ Calif Santa Barbara, Santa Barbara, CA 93106 USA
[2] King Abdullah Univ Sci & Technol, Thuwal, Saudi Arabia
关键词
Generalization error; Binary Classification; Overparameterization; Max-margin; Asymptotics;
D O I
10.1109/icassp40776.2020.9053524
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We consider a model for logistic regression where only a subset of features of size p is used for training a linear classifier over n training samples. The classifier is obtained by running gradient-descent (GD) on the logistic-loss. For this model, we investigate the dependence of the classification error on the overparameterization ratio kappa = p/n. First, building on known deterministic results on convergence properties of the GD, we uncover a phase-transition phenomenon for the case of Gaussian features: the classification error of GD is the same as that of the maximum-likelihood (ML) solution when kappa < kappa(star), and that of the max-margin (SVM) solution when kappa < kappa(star). Next, using the convex Gaussian min-max theorem (CGMT), we sharply characterize the performance of both the ML and SVM solutions. Combining these results, we obtain curves that explicitly characterize the test error of GD for varying values of kappa. The numerical results validate the theoretical predictions and unveil "double-descent" phenomena that complement similar recent observations in linear regression settings.
引用
收藏
页码:4267 / 4271
页数:5
相关论文
共 50 条
  • [31] THE PHASE TRANSITION FOR THE EXISTENCE OF THE MAXIMUM LIKELIHOOD ESTIMATE IN HIGH-DIMENSIONAL LOGISTIC REGRESSION
    Candes, Emmanuel J.
    Sur, Pragya
    ANNALS OF STATISTICS, 2020, 48 (01): : 27 - 42
  • [32] A systematic review on model selection in high-dimensional regression
    Lee, Eun Ryung
    Cho, Jinwoo
    Yu, Kyusang
    JOURNAL OF THE KOREAN STATISTICAL SOCIETY, 2019, 48 (01) : 1 - 12
  • [33] A Model Selection Criterion for High-Dimensional Linear Regression
    Owrang, Arash
    Jansson, Magnus
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2018, 66 (13) : 3436 - 3446
  • [34] A systematic review on model selection in high-dimensional regression
    Eun Ryung Lee
    Jinwoo Cho
    Kyusang Yu
    Journal of the Korean Statistical Society, 2019, 48 : 1 - 12
  • [35] Jackknife model averaging for high-dimensional quantile regression
    Wang, Miaomiao
    Zhang, Xinyu
    Wan, Alan T. K.
    You, Kang
    Zou, Guohua
    BIOMETRICS, 2023, 79 (01) : 178 - 189
  • [36] Model diagnosis for parametric regression in high-dimensional spaces
    Stute, W.
    Xu, W. L.
    Zhu, L. X.
    BIOMETRIKA, 2008, 95 (02) : 451 - 467
  • [37] A Model-Averaging Approach for High-Dimensional Regression
    Ando, Tomohiro
    Li, Ker-Chau
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2014, 109 (505) : 254 - 265
  • [38] SPReM: Sparse Projection Regression Model For High-Dimensional Linear Regression
    Sun, Qiang
    Zhu, Hongtu
    Liu, Yufeng
    Ibrahim, Joseph G.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2015, 110 (509) : 289 - 302
  • [39] Double and Single Descent in Causal Inference with an Application to High-Dimensional Synthetic Control
    Spiess, Jann
    Imbens, Guido
    Venugopal, Amar
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [40] Regression on High-dimensional Inputs
    Kuleshov, Alexander
    Bernstein, Alexander
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2016, : 732 - 739