Dynamic QoS Prediction With Intelligent Route Estimation Via Inverse Reinforcement Learning

被引:2
|
作者
Li, Jiahui [1 ]
Wu, Hao [1 ]
He, Qiang [2 ,3 ]
Zhao, Yiji [4 ]
Wang, Xin [5 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650091, Yunnan, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
[3] Huazhong Univ Sci & Technol, Natl Engn Res Ctr Big Data Technol & Syst, Sch Comp Sci & Technol, Serv Comp Technol & Syst Lab,Cluster & Grid Comp, Wuhan 430074, Peoples R China
[4] Beijing Jiaotong Univ, Sch Comp Sci, Beijing 100044, Peoples R China
[5] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Quality of service; Reinforcement learning; Estimation; Predictive models; Peer-to-peer computing; Network topology; Heuristic algorithms; Deep learning; inverse reinforcement learning; QoS prediction; reinforcement learning; route estimation; LOCATION; RECOMMENDATION; FACTORIZATION; ALGORITHMS; MODEL;
D O I
10.1109/TSC.2023.3342481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dynamic quality of service (QoS) measurement is crucial for discovering services and developing online service systems. Collaborative filtering-based approaches perform dynamic QoS prediction by incorporating temporal information only but never consider the dynamic network environment and suffer from poor performance. Considering different service invocation routes directly reflect the dynamic environment and further lead to QoS fluctuations, we coin the problem of Dynamic QoS Prediction (DQP) with Intelligent Route Estimation (IRE) and propose a novel framework named IRE4DQP. Under the IRE4DQP framework, the dynamic environment is captured by Network Status Representation, and the IRE is modeled as a Markov decision process and implemented by a deep learning agent. After that, the DQP is achieved by a specific neural model with the estimated route as input. Through collaborative training with reinforcement and inverse reinforcement learning, eventually, based on the updated representations of the network status, IRE learns an optimal route policy that matches well with observed QoS values, and DQP achieves accurate predictions. Experimental results demonstrate that IRE4DQP outperforms SOTA methods on the accuracy of response-time prediction by 5.79-31.34% in MAE, by 1.29-20.18% in RMSE, and by 4.43-27.73% in NMAE and with a success rate of nearly 45% on finding routes.
引用
收藏
页码:509 / 523
页数:15
相关论文
共 50 条
  • [21] Inverse Reinforcement Learning Using Dynamic Policy Programming
    Uchibe, Eiji
    Doya, Kenji
    FOUTH JOINT IEEE INTERNATIONAL CONFERENCES ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (IEEE ICDL-EPIROB 2014), 2014, : 222 - 228
  • [22] Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior
    Ashwood, Zoe C.
    Jha, Aditi
    Pillow, JonathanW.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [23] Dynamic multiobjective optimization driven by inverse reinforcement learning
    Zou, Fei
    Yen, Gary G.
    Zhao, Chen
    INFORMATION SCIENCES, 2021, 575 : 468 - 484
  • [24] Inverse Reinforcement Learning via Deep Gaussian Process
    Jin, Ming
    Damianou, Andreas
    Abbeel, Pieter
    Spanos, Costas
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017), 2017,
  • [25] Understanding Sequential Decisions via Inverse Reinforcement Learning
    Liu, Siyuan
    Araujo, Miguel
    Brunskill, Emma
    Rossetti, Rosaldo
    Barros, Joao
    Krishnan, Ramayya
    2013 IEEE 14TH INTERNATIONAL CONFERENCE ON MOBILE DATA MANAGEMENT (MDM 2013), VOL 1, 2013, : 177 - 186
  • [26] IQoR: An Intelligent QoS-aware Routing Mechanism with Deep Reinforcement Learning
    Cao, Yuanyuan
    Dai, Bin
    Mo, Yijun
    Xu, Yang
    PROCEEDINGS OF THE 2020 IEEE 45TH CONFERENCE ON LOCAL COMPUTER NETWORKS (LCN 2020), 2020, : 329 - 332
  • [27] Intelligent Inventory Control via Ruminative Reinforcement Learning
    Katanyukul, Tatpong
    Chong, Edwin K. P.
    JOURNAL OF APPLIED MATHEMATICS, 2014,
  • [28] Intelligent Decision-Making of Scheduling for Dynamic Permutation Flowshop via Deep Reinforcement Learning
    Yang, Shengluo
    Xu, Zhigang
    Wang, Junyi
    SENSORS, 2021, 21 (03) : 1 - 21
  • [29] An Inverse Reinforcement Learning Method to Infer Reward Function of Intelligent Jammer
    Fan, Youlin
    Jiu, Bo
    Pu, Wenqiang
    Li, Kang
    Zhang, Yu
    Liu, Hongwei
    Proceedings of the IEEE Radar Conference, 2023,
  • [30] Inverse Reinforcement Learning: A New Framework to Mitigate an Intelligent Backoff Attack
    Parras, Juan
    Almodovar, Alejandro
    Apellaniz, Patricia A.
    Zazo, Santiago
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (24): : 24790 - 24799