A Two-Stage Deep Reinforcement Learning Framework for MEC-Enabled Adaptive 360-Degree Video Streaming

被引:0
|
作者
Bi, Suzhi [1 ]
Chen, Haoguo [1 ]
Li, Xian [1 ]
Wang, Shuoyao [1 ]
Wu, Yuan [2 ]
Qian, Liping [3 ]
机构
[1] Shenzhen Univ, Coll Elect & Informat Engn, State Key Lab Radio Frequency Heterogeneous Integ, Shenzhen 518060, Peoples R China
[2] Univ Macau, Dept Comp & Informat Sci, State Key Lab Internet Things Smart City, Taipa, Macao, Peoples R China
[3] Zhejiang Univ Technol, Coll Informat Engn, Hangzhou, Zhejiang, Peoples R China
关键词
Streaming media; Bit rate; Quality of experience; Wireless communication; Real-time systems; Resists; Accuracy; Adaptive streaming; multi-access edge computing (MEC); quality of experience (QoE); deep reinforcement learning; RATE ADAPTATION; PREDICTION; COMMUNICATION;
D O I
10.1109/TMC.2024.3443200
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The emerging multi-access edge computing (MEC) technology effectively enhances the wireless streaming performance of 360-degree videos. By connecting a user's head-mounted device (HMD) to a smart MEC platform, the edge server (ES) can efficiently perform adaptive tile-based video streaming to improve the user's viewing experience. Under constrained wireless channel capacity, the ES can predict the user's field of view (FoV) and transmit to the HMD high-resolution video tiles only within the predicted FoV. In practice, the video streaming performance is challenged by the random FoV prediction error and wireless channel fading effects. For this, we propose in this paper a novel two-stage adaptive 360-degree video streaming scheme that maximizes the user's quality of experience (QoE) to attain stable and high-resolution video playback. Specifically, we divide the video file into groups of pictures (GOPs) of fixed playback interval, where each GOP consists of a number of video frames. At the beginning of each GOP (i.e., the inter-GOP stage), the ES predicts the FoV of the next GOP and allocates an encoding bitrate for transmitting (precaching) the video tiles within the predicted FoV. Then, during the real-time video playback of the current GOP (i.e., the intra-GOP stage), the ES observes the user's true FoV of each frame and transmits the missing tiles to compensate for the FoV prediction errors. To maximize the user's QoE under random variations of FoV and wireless channel, we propose a double-agent deep reinforcement learning framework, where the two agents operate in different time scales to decide the bitrates of inter- and intra-GOP stages, respectively. Experiments based on real-world measurements show that the proposed scheme can effectively mitigate FoV prediction errors and maintain stable QoE performance under different scenarios, achieving over 22.1% higher QoE than some representative benchmark methods.
引用
收藏
页码:14313 / 14329
页数:17
相关论文
共 50 条
  • [1] Deep Curriculum Reinforcement Learning for Adaptive 360 Video Streaming With Two-Stage Training
    Xie, Yuhong
    Zhang, Yuan
    Lin, Tao
    IEEE TRANSACTIONS ON BROADCASTING, 2024, 70 (02) : 441 - 452
  • [2] Perceptual Quality Aware Adaptive 360-Degree Video Streaming with Deep Reinforcement Learning
    Feng, Qingxuan
    Yang, Peng
    Lyu, Feng
    Yu, Li
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 1190 - 1195
  • [3] DRL360: 360-degree Video Streaming with Deep Reinforcement Learning
    Zhang, Yuanxing
    Zhao, Pengyu
    Bian, Kaigui
    Liu, Yunxin
    Song, Lingyang
    Li, Xiaoming
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2019), 2019, : 1252 - 1260
  • [4] DEEP REINFORCEMENT LEARNING-BASED RATE ADAPTATION FOR ADAPTIVE 360-DEGREE VIDEO STREAMING
    Kan, Nuowen
    Zou, Junni
    Tang, Kexin
    Li, Chenglin
    Liu, Ning
    Xiong, Hongkai
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 4030 - 4034
  • [5] Adaptive Streaming of 360-Degree Videos with Reinforcement Learning
    Park, Sohee
    Hoai, Minh
    Bhattacharya, Arani
    Das, Samir R.
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1838 - 1847
  • [6] Deep Reinforcement Learning Based Adaptive 360-degree Video Streaming with Field of View Joint Prediction
    Zhang, Yuanhong
    Wang, Zhiwen
    Liu, Junquan
    Du, Haipeng
    Zheng, Qinghua
    Zhang, Weizhan
    2022 27TH IEEE SYMPOSIUM ON COMPUTERS AND COMMUNICATIONS (IEEE ISCC 2022), 2022,
  • [7] Reinforcement Learning Based Rate Adaptation for 360-Degree Video Streaming
    Jiang, Zhiqian
    Zhang, Xu
    Xu, Yiling
    Ma, Zhan
    Sun, Jun
    Zhang, Yunfei
    IEEE TRANSACTIONS ON BROADCASTING, 2021, 67 (02) : 409 - 423
  • [8] Double-Agent Deep Reinforcement Learning for Adaptive 360-Degree Video Streaming in Mobile Edge Computing Networks
    Bi, Suzhi
    Chen, Haoguo
    Li, Xian
    Wang, Shuoyao
    Lin, Xiao-Hui
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS, INFOCOM WKSHPS 2024, 2024,
  • [9] PARIMA: Viewport Adaptive 360-Degree Video Streaming
    Chopra, Lovish
    Chakraborty, Sarthak
    Mondal, Abhijit
    Chakraborty, Sandip
    PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, : 2379 - 2391
  • [10] RAPT360: Reinforcement Learning-Based Rate Adaptation for 360-Degree Video Streaming With Adaptive Prediction and Tiling
    Kan, Nuowen
    Zou, Junni
    Li, Chenglin
    Dai, Wenrui
    Xiong, Hongkai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (03) : 1607 - 1623