Real-time power optimization based on Q-learning algorithm for direct methanol fuel cell system

被引:1
|
作者
Chi, Xuncheng [1 ]
Chen, Fengxiang [1 ]
Zhai, Shuang [2 ]
Hu, Zhe [2 ]
Zhou, Su [3 ]
Wei, Wei [4 ]
机构
[1] Tongji Univ, Sch Automot Studies, Shanghai, Peoples R China
[2] Shanghai Refire Technol Co Ltd, Shanghai, Peoples R China
[3] Shanghai Zhongqiao Vocat & Tech Univ, Shanghai, Peoples R China
[4] CAS &M Zhangjiagang New Energy Technol Co Ltd, Zhangjiagang, Peoples R China
基金
中国国家自然科学基金;
关键词
Direct methanol fuel cell (DMFC) system; Real-time power optimization; Methanol supply control; Reinforcement learning; Q -learning algorithm; MASS-TRANSPORT MODEL; NUMERICAL-MODEL; PERFORMANCE; DMFC;
D O I
10.1016/j.ijhydene.2024.09.084
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
Efficient real-time power optimization of direct methanol fuel cell (DMFC) system is crucial for enhancing its performance and reliability. The power of DMFC system is mainly affected by stack temperature and circulating methanol concentration. However, the methanol concentration cannot be directly measured using reliable sensors, which poses a challenge for the real-time power optimization. To address this issue, this paper investigates the operating mechanism of DMFC system and establishes a system power model. Based on the established model, reinforcement learning using Q-learning algorithm is proposed to control methanol supply to optimize DMFC system power under varying operating conditions. This algorithm is simple, easy to implement, and does not rely on methanol concentration measurements. To validate the effectiveness of the proposed algorithm, simulation comparisons between the proposed method and the traditional perturbation and observation (P&O) algorithm are implemented under different operating conditions. The results show that proposed power optimization based on Q-learning algorithm improves net power by 1% and eliminates the fluctuation of methanol supply caused by P&O. For practical implementation considerations and real-time requirements of the algorithm, hardware-in-the-loop (HIL) experiments are conducted. The experiment results demonstrate that the proposed methods optimize net power under different operating conditions. Additionally, in terms of model accuracy, the experimental results are well matched with the simulation. Moreover, under varying load condition, compared with P&O, proposed power optimization based on Q-learning algorithm reduces root mean square error (RMSE) from 7.271% to 2.996% and mean absolute error (MAE) from 5.036% to 0.331%.
引用
收藏
页码:1241 / 1253
页数:13
相关论文
共 50 条
  • [21] A Power Management System for Direct Methanol Fuel Cell
    Song, H. J.
    Lee, S. J.
    Yoo, E. J.
    Park, H. J.
    Noh, M. G.
    Park, Y. W.
    TENCON 2012 - 2012 IEEE REGION 10 CONFERENCE: SUSTAINABLE DEVELOPMENT THROUGH HUMANITARIAN TECHNOLOGY, 2012,
  • [22] Power management of a direct methanol fuel cell system
    Jiang, Rongzhong
    Chu, Deryn
    JOURNAL OF POWER SOURCES, 2006, 161 (02) : 1192 - 1197
  • [23] Q-RTS: a real-time swarm intelligence based on multi- agent Q-learning
    Matta, M.
    Cardarilli, G. C.
    Di Nunzio, L.
    Fazzolari, R.
    Giardino, D.
    Re, M.
    Silvestri, F.
    Spano, S.
    ELECTRONICS LETTERS, 2019, 55 (10) : 589 - +
  • [24] Coverage Path Planning Optimization Based on Q-Learning Algorithm
    Piardi, Luis
    Lima, Jose
    Pereira, Ana, I
    Costa, Paulo
    INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS (ICNAAM-2018), 2019, 2116
  • [25] Research on real-time scheduling optimization technology of power system based on deep learning
    Lu, Min
    Jiang, Yicheng
    Wang, Jin
    Zhu, Jianping
    Applied Mathematics and Nonlinear Sciences, 2024, 9 (01)
  • [26] Multi-objective optimization of a direct methanol fuel cell system using a genetic-based algorithm
    Mert, Suha Orcun
    Ozcelik, Zehra
    INTERNATIONAL JOURNAL OF ENERGY RESEARCH, 2013, 37 (10) : 1256 - 1264
  • [27] Real-time Automated Guided vehicles scheduling with Markov Decision Process and Double Q-Learning algorithm
    Sagar, Kiran, V
    Jerald, J.
    MATERIALS TODAY-PROCEEDINGS, 2022, 64 : 279 - 284
  • [28] Energy-Efficient Scheduling for Real-Time Systems Based on Deep Q-Learning Mode
    Zhang, Qingchen
    Lin, Man
    Yang, Laurence T.
    Chen, Zhikui
    Li, Peng
    IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING, 2019, 4 (01): : 132 - 141
  • [29] Real-Time Adjustment Method for Metro Systems with Train Delays Based on Improved Q-Learning
    Hu, Yushen
    Li, Wei
    Luo, Qin
    APPLIED SCIENCES-BASEL, 2024, 14 (04):
  • [30] Autonomous Navigation based on a Q-learning algorithm for a Robot in a Real Environment
    Strauss, Clement
    Sahin, Ferat
    2008 IEEE INTERNATIONAL CONFERENCE ON SYSTEM OF SYSTEMS ENGINEERING (SOSE), 2008, : 361 - 365