Dynamic programming with meta-reinforcement learning: a novel approach for multi-objective optimization

被引:1
|
作者
Wang, Qi [1 ]
Zhang, Chengwei [1 ]
Hu, Bin [2 ]
机构
[1] Dalian Maritime Univ, Informat Sci & Technol Coll, Dalian, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou, Peoples R China
关键词
Combinatorial optimization; Meta-learning; Reinforcement learning; Dynamic programming; PERSON REIDENTIFICATION;
D O I
10.1007/s40747-024-01469-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-objective optimization (MOO) endeavors to identify optimal solutions from a finite array of possibilities. In recent years, deep reinforcement learning (RL) has exhibited promise through its well-crafted heuristics in tackling NP-hard combinatorial optimization (CO) problems. Nonetheless, current methodologies grapple with two key challenges: (1) They primarily concentrate on single-objective optimization quandaries, rendering them less adaptable to the more prevalent MOO scenarios encountered in real-world applications. (2) These approaches furnish an approximate solution by imbibing heuristics, lacking a systematic means to enhance or substantiate optimality. Given these challenges, this study introduces an overarching hybrid strategy, dynamic programming with meta-reinforcement learning (DPML), to resolve MOO predicaments. The approach melds meta-learning into an RL framework, addressing multiple subproblems inherent to MOO. Furthermore, the precision of solutions is elevated by endowing exact dynamic programming with the prowess of meta-graph neural networks. Empirical results substantiate the supremacy of our methodology over previous RL and heuristics approaches, bridging the chasm between theoretical underpinnings and real-world applicability within this domain.
引用
收藏
页码:5743 / 5758
页数:16
相关论文
共 50 条
  • [21] Multi-workflow dynamic scheduling in product design: A generalizable approach based on meta-reinforcement learning
    Chen, Zhen
    Zhang, Lin
    Cai, Wentong
    Laili, Yuanjun
    Wang, Xiaohan
    Wang, Fei
    Wang, Huijuan
    JOURNAL OF MANUFACTURING SYSTEMS, 2025, 79 : 334 - 346
  • [22] Prediction Guided Meta-Learning for Multi-Objective Reinforcement Learning
    Liu, Fei-Yu
    Qian, Chao
    2021 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC 2021), 2021, : 2171 - 2178
  • [23] A context-based meta-reinforcement learning approach to efficient hyperparameter optimization
    Liu, Xiyuan
    Wu, Jia
    Chen, Senpeng
    NEUROCOMPUTING, 2022, 478 : 89 - 103
  • [24] Dynamic Channel Access via Meta-Reinforcement Learning
    Lu, Ziyang
    Gursoy, M. Cenk
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [25] Multi-objective reinforcement learning-based approach for pressurized water reactor optimization
    Seurin, Paul
    Shirvan, Koroush
    ANNALS OF NUCLEAR ENERGY, 2024, 205
  • [26] Reinforcement learning with multi-objective optimization in targeted drug design
    Abbasi, M.
    EUROPEAN JOURNAL OF CLINICAL INVESTIGATION, 2021, 51 : 102 - 103
  • [27] A Multi-objective Reinforcement Learning Solution for Handover Optimization in URLLC
    Arnaz, Azadeh
    Lipman, Justin
    Abolhasan, Mehran
    2023 28TH ASIA PACIFIC CONFERENCE ON COMMUNICATIONS, APCC 2023, 2023, : 68 - 74
  • [28] Multi-Objective Optimization Using Adaptive Distributed Reinforcement Learning
    Tan, Jing
    Khalili, Ramin
    Karl, Holger
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (09) : 10777 - 10789
  • [29] An Improved Multi-objective Optimization Algorithm Based on Reinforcement Learning
    Liu, Jun
    Zhou, Yi
    Qiu, Yimin
    Li, Zhongfeng
    ADVANCES IN SWARM INTELLIGENCE, ICSI 2022, PT I, 2022, : 501 - 513
  • [30] Optimization of DEM parameters using multi-objective reinforcement learning
    Westbrink, Fabian
    Elbel, Alexander
    Schwung, Andreas
    Ding, Steven X.
    POWDER TECHNOLOGY, 2021, 379 : 602 - 616