DEN-DQL: Quick Convergent Deep Q-Learning with Double Exploration Networks for News Recommendation

被引:0
|
作者
Song, Zhanghan [1 ]
Zhang, Dian [1 ]
Shi, Xiaochuan [1 ]
Li, Wei [2 ]
Ma, Chao [1 ]
Wu, Libing [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan, Peoples R China
[2] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Wuxi, Jiangsu, Peoples R China
关键词
Reinforcement Learning; Deep Q-Learning; News Recommendation; Double Exploration Networks;
D O I
10.1109/IJCNN52387.2021.9533818
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the dynamic characteristics of news and user preferences, personalized recommendation is a challenging problem. Traditional recommendation methods simply focus on current reward, which just recommend items to maximize the number of current clicks. And this may reduce users' interest in similar items. Although the news recommendation framework based on deep reinforcement learning preciously proposed (i.e, DRL, based on deep Q-learning) has the advantages of focusing on future total rewards and dynamic interactive recommendation, it has two issues. First, its exploration method is slow to converge, which may bring new users a bad experience. Second, it is hard to train on off-line data set because the reward is difficult to be determined. In order to address the aforementioned issues, we propose a framework named DEN-DQL for news recommendation based on deep Q-learning with double exploration networks. Also, we develop a new method to calculate rewards and use an off-line data set to simulate the online news clicking environment to train DEN-DQL. Then, the well trained DEN-DQL is tested in the online environment of the same data set, which demonstrates at least 10% improvement of the proposed DEN-DQL.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100
  • [2] Adaptive Learning Recommendation Strategy Based on Deep Q-learning
    Tan, Chunxi
    Han, Ruijian
    Ye, Rougang
    Chen, Kani
    APPLIED PSYCHOLOGICAL MEASUREMENT, 2020, 44 (04) : 251 - 266
  • [3] ADRP-DQL: An adaptive distributed routing protocol for underwater acoustic sensor networks using deep Q-learning
    Mohanraju, Adi Surendra M.
    Lokam, Anjaneyulu
    AD HOC NETWORKS, 2025, 167
  • [4] Experimental Research on Avoidance Obstacle Control for Mobile Robots Using Q-Learning (QL) and Deep Q-Learning (DQL) Algorithms in Dynamic Environments
    Ha, Vo Thanh
    Vinh, Vo Quang
    ACTUATORS, 2024, 13 (01)
  • [5] Double Deep Q-Learning Based Channel Estimation for Industrial Wireless Networks
    Bhardwaj, Sanjay
    Lee, Jae-Min
    Kim, Dong-Seong
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1318 - 1320
  • [6] Maximizing Opinion Polarization Using Double Deep Q-Learning in Social Networks
    Zareer, Mohamed N.
    Selmic, Rastko R.
    IEEE ACCESS, 2025, 13 : 57398 - 57412
  • [7] QLP: Deep Q-Learning for Pruning Deep Neural Networks
    Camci, Efe
    Gupta, Manas
    Wu, Min
    Lin, Jie
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 6488 - 6501
  • [8] A Distributed Double Deep Q-Learning Method for Object Redundancy Mitigation in Vehicular Networks
    Ghnaya, Imed
    Aniss, Hasnaa
    Ahmed, Toufik
    Mosbah, Mohamed
    2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [9] Multi-Agent Double Deep Q-Learning for Beamforming in mmWave MIMO Networks
    Wang, Xueyuan
    Gursoy, M. Cenk
    2020 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (IEEE PIMRC), 2020,
  • [10] Federated Double Deep Q-learning for Joint Delay and Energy Minimization in IoT networks
    Zarandi, Sheyda
    Tabassum, Hina
    2021 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2021,