Dynamic QoS Prediction With Intelligent Route Estimation Via Inverse Reinforcement Learning

被引:2
|
作者
Li, Jiahui [1 ]
Wu, Hao [1 ]
He, Qiang [2 ,3 ]
Zhao, Yiji [4 ]
Wang, Xin [5 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650091, Yunnan, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
[3] Huazhong Univ Sci & Technol, Natl Engn Res Ctr Big Data Technol & Syst, Sch Comp Sci & Technol, Serv Comp Technol & Syst Lab,Cluster & Grid Comp, Wuhan 430074, Peoples R China
[4] Beijing Jiaotong Univ, Sch Comp Sci, Beijing 100044, Peoples R China
[5] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Quality of service; Reinforcement learning; Estimation; Predictive models; Peer-to-peer computing; Network topology; Heuristic algorithms; Deep learning; inverse reinforcement learning; QoS prediction; reinforcement learning; route estimation; LOCATION; RECOMMENDATION; FACTORIZATION; ALGORITHMS; MODEL;
D O I
10.1109/TSC.2023.3342481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dynamic quality of service (QoS) measurement is crucial for discovering services and developing online service systems. Collaborative filtering-based approaches perform dynamic QoS prediction by incorporating temporal information only but never consider the dynamic network environment and suffer from poor performance. Considering different service invocation routes directly reflect the dynamic environment and further lead to QoS fluctuations, we coin the problem of Dynamic QoS Prediction (DQP) with Intelligent Route Estimation (IRE) and propose a novel framework named IRE4DQP. Under the IRE4DQP framework, the dynamic environment is captured by Network Status Representation, and the IRE is modeled as a Markov decision process and implemented by a deep learning agent. After that, the DQP is achieved by a specific neural model with the estimated route as input. Through collaborative training with reinforcement and inverse reinforcement learning, eventually, based on the updated representations of the network status, IRE learns an optimal route policy that matches well with observed QoS values, and DQP achieves accurate predictions. Experimental results demonstrate that IRE4DQP outperforms SOTA methods on the accuracy of response-time prediction by 5.79-31.34% in MAE, by 1.29-20.18% in RMSE, and by 4.43-27.73% in NMAE and with a success rate of nearly 45% on finding routes.
引用
收藏
页码:509 / 523
页数:15
相关论文
共 50 条
  • [1] Learning Tasks in Intelligent Environments via Inverse Reinforcement Learning
    Shah, Syed Ihtesham Hussain
    Coronato, Antonio
    2021 17TH INTERNATIONAL CONFERENCE ON INTELLIGENT ENVIRONMENTS (IE), 2021,
  • [2] Estimation of personal driving style via deep inverse reinforcement learning
    Kishikawa, Daiko
    Arai, Sachiyo
    ARTIFICIAL LIFE AND ROBOTICS, 2021, 26 (03) : 338 - 346
  • [3] Estimation of personal driving style via deep inverse reinforcement learning
    Daiko Kishikawa
    Sachiyo Arai
    Artificial Life and Robotics, 2021, 26 : 338 - 346
  • [4] Probabilistic Prediction of Interactive Driving Behavior via Hierarchical Inverse Reinforcement Learning
    Sun, Liting
    Zhan, Wei
    Tomizuka, Masayoshi
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2111 - 2117
  • [5] Future Trajectory Prediction via RNN and Maximum Margin Inverse Reinforcement Learning
    Choi, Dooseop
    An, Taeg-Hyun
    Ahn, Kyounghwan
    Choi, Jeongdan
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 125 - 130
  • [6] Active Learning for Reward Estimation in Inverse Reinforcement Learning
    Lopes, Manuel
    Melo, Francisco
    Montesano, Luis
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT II, 2009, 5782 : 31 - +
  • [7] Inverse Reinforcement Learning for Legibility Automation in Intelligent Agents
    Zeng, Buxin
    Zeng, Yifeng
    Pan, Yinghui
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 741 - 746
  • [8] Machining sequence learning via inverse reinforcement learning
    Sugisawa, Yasutomo
    Takasugi, Keigo
    Asakawa, Naoki
    PRECISION ENGINEERING-JOURNAL OF THE INTERNATIONAL SOCIETIES FOR PRECISION ENGINEERING AND NANOTECHNOLOGY, 2022, 73 : 477 - 487
  • [9] Reinforcement learning as a means of dynamic aggregate QoS provisioning
    Akar, N
    Sahin, C
    ARCHITECTURES FOR QUALITY OF SERVICE IN THE INTERNET, 2003, 2698 : 100 - 114
  • [10] Regularising neural networks for future trajectory prediction via inverse reinforcement learning framework
    Choi, Dooseop
    Min, Kyoungwook
    Choi, Jeongdan
    IET COMPUTER VISION, 2020, 14 (05) : 192 - 200