To understand double descent, we need to understand VC theory

被引:3
|
作者
Cherkassky, Vladimir [1 ]
Lee, Eng Hock [1 ]
机构
[1] Univ Minnesota Twin Cities, Dept Elect & Comp Engn, Minneapolis, MN 55455 USA
基金
美国国家科学基金会;
关键词
Double descent; Deep learning; Complexity control; Structural risk minimization; VC-dimension; VC-generalization bounds; NEURAL-NETWORKS; BOUNDS; DIMENSION;
D O I
10.1016/j.neunet.2023.10.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We analyze generalization performance of over-parameterized learning methods for classification, under VC theoretical framework. Recently, practitioners in Deep Learning discovered 'double descent' phenomenon, when large networks can fit perfectly available training data, and at the same time, achieve good generalization for future (test) data. The current consensus view is that VC-theoretical results cannot account for good generalization performance of Deep Learning networks. In contrast, this paper shows that double descent can be explained by VC-theoretical concepts, such as VC-dimension and Structural Risk Minimization. We also present empirical results showing that double descent generalization curves can be accurately modeled using classical VC-generalization bounds. Proposed VC-theoretical analysis enables better understanding of generalization curves for data sets with different statistical characteristics, such as low vs high-dimensional data and noisy data. In addition, we analyze generalization performance of transfer learning using pre-trained Deep Learning networks.
引用
收藏
页码:242 / 256
页数:15
相关论文
共 50 条