Neural Q-learning for solving PDEs

被引:0
|
作者
Cohen, Samuel N. [1 ]
Jiang, Deqing [1 ]
Sirignano, Justin [1 ]
机构
[1] Univ Oxford, Math Inst, Oxford OX2 6GG, England
基金
英国工程与自然科学研究理事会;
关键词
Deep learning; neural networks; high-dimensional PDEs; high-dimensional learning; Q-learning; BOUNDARY-VALUE-PROBLEMS; DIFFERENTIAL-EQUATIONS; APPROXIMATION; NETWORK; ALGORITHM; OPERATORS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Solving high-dimensional partial differential equations (PDEs) is a major challenge in scientific computing. We develop a new numerical method for solving elliptic-type PDEs by adapting the Q-learning algorithm in reinforcement learning. To solve PDEs with Dirichlet boundary condition, our "Q-PDE" algorithm is mesh-free and therefore has the potential to overcome the curse of dimensionality. Using a neural tangent kernel (NTK) approach, we prove that the neural network approximator for the PDE solution, trained with the QPDE algorithm, converges to the trajectory of an infinite-dimensional ordinary differential equation (ODE) as the number of hidden units - infinity. For monotone PDEs (i.e. those given by monotone operators, which may be nonlinear), despite the lack of a spectral gap in the NTK, we then prove that the limit neural network, which satisfies the infinite-dimensional ODE, strongly converges in L2 to the PDE solution as the training time - infinity. More generally, we can prove that any fixed point of the wide-network limit for the Q-PDE algorithm is a solution of the PDE (not necessarily under the monotone condition). The numerical performance of the Q-PDE algorithm is studied for several elliptic PDEs.
引用
收藏
页数:49
相关论文
共 50 条
  • [41] Q-learning automaton
    Qian, F
    Hirata, H
    IEEE/WIC INTERNATIONAL CONFERENCE ON INTELLIGENT AGENT TECHNOLOGY, PROCEEDINGS, 2003, : 432 - 437
  • [42] Periodic Q-Learning
    Lee, Donghwan
    He, Niao
    LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 : 582 - 598
  • [43] Mutual Q-learning
    Reid, Cameron
    Mukhopadhyay, Snehasis
    2020 3RD INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTS (ICCR 2020), 2020, : 128 - 133
  • [44] Robust Q-Learning
    Ertefaie, Ashkan
    McKay, James R.
    Oslin, David
    Strawderman, Robert L.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2021, 116 (533) : 368 - 381
  • [45] Logistic Q-Learning
    Bas-Serrano, Joan
    Curi, Sebastian
    Krause, Andreas
    Neu, Gergely
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [46] Q-learning based on neural network in learning action selection of mobile robot
    Qiao, Junfei
    Hou, Zhanjun
    Ruan, Xiaogang
    2007 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS, VOLS 1-6, 2007, : 263 - 267
  • [47] Enhancing Nash Q-learning and Team Q-learning mechanisms by using bottlenecks
    Ghazanfari, Behzad
    Mozayani, Nasser
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2014, 26 (06) : 2771 - 2783
  • [48] Experiments with Solving Mountain Car Problem Using State Discretization and Q-Learning
    Badica, Amelia
    Badica, Costin
    Ivanovic, Mirjana
    Logofatu, Doina
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2022, PT I, 2022, 13757 : 142 - 155
  • [49] Improved Q-learning algorithm for solving permutation flow shop scheduling problems
    He, Zimiao
    Wang, Kunlan
    Li, Hanxiao
    Song, Hong
    Lin, Zhongjie
    Gao, Kaizhou
    Sadollah, Ali
    IET COLLABORATIVE INTELLIGENT MANUFACTURING, 2022, 4 (01) : 35 - 44
  • [50] Solving the optimal path planning of a mobile robot using improved Q-learning
    Low, Ee Soong
    Ong, Pauline
    Cheah, Kah Chun
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 115 : 143 - 161