Neural Q-learning for solving PDEs

被引:0
|
作者
Cohen, Samuel N. [1 ]
Jiang, Deqing [1 ]
Sirignano, Justin [1 ]
机构
[1] Univ Oxford, Math Inst, Oxford OX2 6GG, England
基金
英国工程与自然科学研究理事会;
关键词
Deep learning; neural networks; high-dimensional PDEs; high-dimensional learning; Q-learning; BOUNDARY-VALUE-PROBLEMS; DIFFERENTIAL-EQUATIONS; APPROXIMATION; NETWORK; ALGORITHM; OPERATORS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Solving high-dimensional partial differential equations (PDEs) is a major challenge in scientific computing. We develop a new numerical method for solving elliptic-type PDEs by adapting the Q-learning algorithm in reinforcement learning. To solve PDEs with Dirichlet boundary condition, our "Q-PDE" algorithm is mesh-free and therefore has the potential to overcome the curse of dimensionality. Using a neural tangent kernel (NTK) approach, we prove that the neural network approximator for the PDE solution, trained with the QPDE algorithm, converges to the trajectory of an infinite-dimensional ordinary differential equation (ODE) as the number of hidden units - infinity. For monotone PDEs (i.e. those given by monotone operators, which may be nonlinear), despite the lack of a spectral gap in the NTK, we then prove that the limit neural network, which satisfies the infinite-dimensional ODE, strongly converges in L2 to the PDE solution as the training time - infinity. More generally, we can prove that any fixed point of the wide-network limit for the Q-PDE algorithm is a solution of the PDE (not necessarily under the monotone condition). The numerical performance of the Q-PDE algorithm is studied for several elliptic PDEs.
引用
收藏
页数:49
相关论文
共 50 条
  • [21] Q-learning and traditional methods on solving the pocket Rubik's cube
    Lyu, Zefeng
    Liu, Zeyu
    Khojandi, Anahita
    Yu, Andrew Junfang
    COMPUTERS & INDUSTRIAL ENGINEERING, 2022, 171
  • [22] Solving a Job Shop Scheduling Problem Using Q-Learning Algorithm
    Belmamoune, Manal Abir
    Ghomri, Latefa
    Yahouni, Zakaria
    12TH INTERNATIONAL WORKSHOP ON SERVICE ORIENTED, HOLONIC AND MULTI-AGENT MANUFACTURING SYSTEMS FOR INDUSTRY OF THE FUTURE, SOHOMA 2022, 2023, 1083 : 196 - 209
  • [23] An Algorithm Based on Q-learning for Solving Frequency Assignment in RFID Systems
    Hu Shengbo
    2008 CHINESE CONTROL AND DECISION CONFERENCE, VOLS 1-11, 2008, : 225 - 229
  • [24] Accelerating Nash Q-Learning with Graphical Game Representation and Equilibrium Solving
    Zhuang, Yunkai
    Chen, Xingguo
    Gao, Yang
    Hu, Yujing
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 939 - 946
  • [25] A Differential Evolution Algorithm with Q-Learning for Solving Engineering Design Problems
    Kizilay, Damla
    Tasgetiren, M. Fatih
    Oztop, Hande
    Kandiller, Levent
    Suganthan, P. N.
    2020 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2020,
  • [26] Constrained Deep Q-Learning Gradually Approaching Ordinary Q-Learning
    Ohnishi, Shota
    Uchibe, Eiji
    Yamaguchi, Yotaro
    Nakanishi, Kosuke
    Yasui, Yuji
    Ishii, Shin
    FRONTIERS IN NEUROROBOTICS, 2019, 13
  • [27] Learning rates for Q-Learning
    Even-Dar, E
    Mansour, Y
    COMPUTATIONAL LEARNING THEORY, PROCEEDINGS, 2001, 2111 : 589 - 604
  • [28] Learning rates for Q-learning
    Even-Dar, E
    Mansour, Y
    JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 5 : 1 - 25
  • [29] Reinforcement Q-Learning and Neural Networks to Acquire Negotiation Behaviors
    Chohra, Amine
    Madani, Kurosh
    Kanzari, Dalel
    NEW CHALLENGES IN APPLIED INTELLIGENCE TECHNOLOGIES, 2008, 134 : 23 - 33
  • [30] Faster Deep Q-learning using Neural Episodic Control
    Nishio, Daichi
    Yamane, Satoshi
    2018 IEEE 42ND ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE (COMPSAC), VOL 1, 2018, : 486 - 491