Enhancing Noisy Binary Search Efficiency through Deep Reinforcement Learning

被引:0
|
作者
Ma, Rui [1 ]
Tao, Yudong [1 ]
Khodeiry, Mohamed M. [2 ]
Alawa, Karam A. [2 ]
Shyu, Mei-Ling [3 ]
Lee, Richard K. [1 ,2 ]
机构
[1] Univ Miami, Dept Elect & Comp Engn, Coral Gables, FL USA
[2] Univ Miami, Bascom Palmer Eye Inst, Miller Sch Med, Miami, FL USA
[3] Univ Missouri Kansas City, Sch Sci & Engn, Kansas City, MO USA
关键词
Deep learning; deep reinforcement learning; noisy binary search;
D O I
10.1109/IRI58017.2023.00033
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Noisy binary search (NBS) aims to find the closest element to a target value within a sorted array through erroneous queries. In an ideal NBS environment where the error rate remains constant, and the costs of all queries are the same, the maximum likelihood estimation (MLE) procedure has been proven to be the optimal decision strategy. However, in some non-ideal NBS problems, both the error rates and the costs are dependent on the queries, and in some cases, finding the optimal decision strategies can be intractable. We propose to use deep reinforcement learning to approximate the optimal decision strategy in the NBS problem, in which an intelligent agent is used to interact with the NBS environment. A dueling double deep Q-network guides the agent to take action at each step, either to generate a query or to stop the search and predict the target value. An optimized policy will be derived by training the network in the NBS environment until convergence. By evaluating our proposed algorithm on a non-ideal NBS environment, visual field test, we show that the performance of our proposed algorithm surpasses baseline visual field testing algorithms by a large margin.
引用
收藏
页码:154 / 159
页数:6
相关论文
共 50 条
  • [31] Enhancing Pokemon VGC Player Performance: Intelligent Agents Through Deep Reinforcement Learning and Neuroevolution
    Rodriguez, Gian
    Villanueva, Edwin
    Baldeon, Johan
    HCI IN GAMES, PT I, HCI-GAMES 2024, 2024, 14730 : 275 - 294
  • [32] Enhancing Explainability of Deep Reinforcement Learning Through Selective Layer-Wise Relevance Propagation
    Huber, Tobias
    Schiller, Dominik
    Andre, Elisabeth
    ADVANCES IN ARTIFICIAL INTELLIGENCE, KI 2019, 2019, 11793 : 188 - 202
  • [33] Training Binary Neural Networks through Learning with Noisy Supervision
    Han, Kai
    Wang, Yunhe
    Xu, Yixing
    Xu, Chunjing
    Wu, Enhua
    Xu, Chang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [34] Using deep reinforcement learning to search reachability properties in systems specified through graph transformation
    Mehrabi, Mohammad Javad
    Rafe, Vahid
    SOFT COMPUTING, 2022, 26 (18) : 9635 - 9663
  • [35] Using deep reinforcement learning to search reachability properties in systems specified through graph transformation
    Mohammad Javad Mehrabi
    Vahid Rafe
    Soft Computing, 2022, 26 : 9635 - 9663
  • [36] Air conditioner component optimum operation point search through a deep reinforcement learning algorithm
    Yoon, Myung-Sup
    Yoon, Won-Sik
    Seo, Myung-Kyo
    Ryu, Seung-Yup
    Lee, Jong-Seok
    2020 20TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2020, : 365 - 372
  • [37] Enhancing E-commerce Product Search through Reinforcement Learning-Powered Query Reformulation
    Agrawal, Sanjay
    Merugu, Srujana
    Sembium, Vivek
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 4488 - 4494
  • [38] Noisy binary search and its applications
    Karp, Richard M.
    Kleinberg, Robert
    PROCEEDINGS OF THE EIGHTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, 2007, : 881 - 890
  • [39] Learn to Steer through Deep Reinforcement Learning
    Wu, Keyu
    Esfahani, Mahdi Abolfazli
    Yuan, Shenghai
    Wang, Han
    SENSORS, 2018, 18 (11)
  • [40] Autonomous exploration through deep reinforcement learning
    Yan, Xiangda
    Huang, Jie
    He, Keyan
    Hong, Huajie
    Xu, Dasheng
    INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2023, 50 (05): : 793 - 803