Over-parameterized Deep Nonparametric Regression for Dependent Data with Its Applications to Reinforcement Learning

被引:0
|
作者
Feng, Xingdong [1 ]
Jiao, Yuling [2 ]
Kang, Lican [3 ]
Zhang, Baqun [1 ]
Zhou, Fan [1 ]
机构
[1] Shanghai Univ Finance & Econ, Sch Stat & Management, Shanghai, Peoples R China
[2] Wuhan Univ, Hubei Key Lab Computat Sci, Sch Math & Stat, Wuhan, Peoples R China
[3] Wuhan Univ, Sch Math & Stat, Wuhan, Peoples R China
基金
中国国家自然科学基金; 上海市科技启明星计划;
关键词
Deep reinforcement learning; Low-dimensional Riemannian manifold; Penalized regression; beta-mixing; NEURAL-NETWORKS; GENERALIZATION ERROR; POLICY ITERATION; APPROXIMATION; BOUNDS; CONVERGENCE; SYSTEMS; RATES; GAME;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we provide statistical guarantees for over-parameterized deep nonparametric regression in the presence of dependent data. By decomposing the error, we establish non-asymptotic error bounds for deep estimation, which is achieved by effectively balancing the approximation and generalization errors. We have derived an approximation result for Holder functions with constrained weights. Additionally, the generalization error is bounded by the weight norm, allowing for a neural network parameter number that is much larger than the training sample size. Furthermore, we address the issue of the curse of dimensionality by assuming that the samples originate from distributions with low intrinsic dimensions. Under this assumption, we are able to overcome the challenges posed by high-dimensional spaces. By incorporating an additional error propagation mechanism, we derive oracle inequalities for the over-parameterized deep fitted Q-iteration.
引用
收藏
页数:40
相关论文
共 50 条