Dynamic programming with meta-reinforcement learning: a novel approach for multi-objective optimization

被引:1
|
作者
Wang, Qi [1 ]
Zhang, Chengwei [1 ]
Hu, Bin [2 ]
机构
[1] Dalian Maritime Univ, Informat Sci & Technol Coll, Dalian, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou, Peoples R China
关键词
Combinatorial optimization; Meta-learning; Reinforcement learning; Dynamic programming; PERSON REIDENTIFICATION;
D O I
10.1007/s40747-024-01469-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-objective optimization (MOO) endeavors to identify optimal solutions from a finite array of possibilities. In recent years, deep reinforcement learning (RL) has exhibited promise through its well-crafted heuristics in tackling NP-hard combinatorial optimization (CO) problems. Nonetheless, current methodologies grapple with two key challenges: (1) They primarily concentrate on single-objective optimization quandaries, rendering them less adaptable to the more prevalent MOO scenarios encountered in real-world applications. (2) These approaches furnish an approximate solution by imbibing heuristics, lacking a systematic means to enhance or substantiate optimality. Given these challenges, this study introduces an overarching hybrid strategy, dynamic programming with meta-reinforcement learning (DPML), to resolve MOO predicaments. The approach melds meta-learning into an RL framework, addressing multiple subproblems inherent to MOO. Furthermore, the precision of solutions is elevated by endowing exact dynamic programming with the prowess of meta-graph neural networks. Empirical results substantiate the supremacy of our methodology over previous RL and heuristics approaches, bridging the chasm between theoretical underpinnings and real-world applicability within this domain.
引用
收藏
页码:5743 / 5758
页数:16
相关论文
共 50 条
  • [31] Multi-objective Optimization of Notifications Using Offline Reinforcement Learning
    Prabhakar, Prakruthi
    Yuan, Yiping
    Yang, Guangyu
    Sun, Wensheng
    Muralidharan, Ajith
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 3752 - 3760
  • [32] Hypernetworks in Meta-Reinforcement Learning
    Beck, Jacob
    Jackson, Matthew
    Vuorio, Risto
    Whiteson, Shimon
    CONFERENCE ON ROBOT LEARNING, VOL 205, 2022, 205 : 1478 - 1487
  • [33] Meta-learning-based multi-objective PSO model for dynamic scheduling optimization
    Lv, Zheng
    Liao, Zherun
    Liu, Ying
    Zhao, Jun
    ENERGY REPORTS, 2023, 9 : 1227 - 1236
  • [34] Meta-learning-based multi-objective PSO model for dynamic scheduling optimization
    Lv, Zheng
    Liao, Zherun
    Liu, Ying
    Zhao, Jun
    ENERGY REPORTS, 2023, 9 : 1227 - 1236
  • [35] Scalarized Multi-Objective Reinforcement Learning: Novel Design Techniques
    Van Moffaert, Kristof
    Drugan, Madalina M.
    Nowe, Ann
    PROCEEDINGS OF THE 2013 IEEE SYMPOSIUM ON ADAPTIVE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING (ADPRL), 2013, : 191 - 199
  • [36] Multi-objective ω-Regular Reinforcement Learning
    Hahn, Ernst Moritz
    Perez, Mateo
    Schewe, Sven
    Somenzi, Fabio
    Trivedi, Ashutosh
    Wojtczak, Dominik
    FORMAL ASPECTS OF COMPUTING, 2023, 35 (02)
  • [37] Federated multi-objective reinforcement learning
    Zhao, Fangyuan
    Ren, Xuebin
    Yang, Shusen
    Zhao, Peng
    Zhang, Rui
    Xu, Xinxin
    INFORMATION SCIENCES, 2023, 624 : 811 - 832
  • [38] An example-based learning approach to multi-objective programming
    Amano, M
    Okano, H
    MULTI-OBJECTIVE PROGRAMMING AND GOAL PROGRAMMING, 2003, : 87 - 92
  • [39] Multi-Objective Optimisation by Reinforcement Learning
    Liao, H. L.
    Wu, Q. H.
    2010 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2010,
  • [40] Multi-Objective Meta Learning
    Ye, Feiyang
    Lin, Baijiong
    Yue, Zhixiong
    Guo, Pengxin
    Xiao, Qiao
    Zhang, Yu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34