Near-optimal Bayesian active learning with correlated and noisy tests

被引:0
|
作者
Chen, Yuxin [1 ]
Hassani, S. Hamed [2 ]
Krause, Andreas [3 ]
机构
[1] CALTECH, Pasadena, CA 91125 USA
[2] Univ Penn, Philadelphia, PA 19104 USA
[3] Swiss Fed Inst Technol, Zurich, Switzerland
来源
ELECTRONIC JOURNAL OF STATISTICS | 2017年 / 11卷 / 02期
基金
欧洲研究理事会;
关键词
Bayesian active learning; information gathering; decision making; noisy observation; approximation algorithms;
D O I
10.1214/17-EJS1336SI
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider the Bayesian active learning and experimental design problem, where the goal is to learn the value of some unknown target variable through a sequence of informative, noisy tests. In contrast to prior work, we focus on the challenging, yet practically relevant setting where test outcomes can be conditionally dependent given the hidden target variable. Under such assumptions, common heuristics, such as greedily performing tests that maximize the reduction in uncertainty of the target, often perform poorly. We propose ECED, a novel, efficient active learning algorithm, and prove strong theoretical guarantees that hold with correlated, noisy tests. Rather than directly optimizing the prediction error, at each step, ECED picks the test that maximizes the gain in a surrogate objective, which takes into account the dependencies between tests. Our analysis relies on an information-theoretic auxiliary function to track the progress of ECED, and utilizes adaptive submodularity to attain the approximation bound. We demonstrate strong empirical performance of ECED on three problem instances, including a Bayesian experimental design task intended to distinguish among economic theories of how people make risky decisions, an active preference learning task via pairwise comparisons, and a third application on pool-based active learning.
引用
收藏
页码:4969 / 5017
页数:49
相关论文
共 50 条
  • [41] Near-Optimal No-Regret Learning for Correlated Equilibria in Multi-player General-Sum Games
    Anagnostides, Ioannis
    Daskalakis, Constantinos
    Farina, Gabriele
    Fishelson, Maxwell
    Golowich, Noah
    Sandholm, Tuomas
    PROCEEDINGS OF THE 54TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING (STOC '22), 2022, : 736 - 749
  • [42] Deriving a Near-optimal Power Management Policy Using Model-Free Reinforcement Learning and Bayesian Classification
    Wang, Yanzhi
    Xie, Qing
    Ammari, Ahmed
    Pedram, Massoud
    PROCEEDINGS OF THE 48TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2011, : 41 - 46
  • [43] Near-Optimal Design of Safe Output-Feedback Controllers From Noisy Data
    Furieri, Luca
    Guo, Baiwei
    Martin, Andrea
    Ferrari-Trecate, Giancarlo
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (05) : 2699 - 2714
  • [44] Comparative Synthesis: Learning Near-Optimal Network Designs by Query
    Wang, Yanjun
    Li, Zixuan
    Jiang, Chuan
    Qiu, Xiaokang
    Rao, Sanjay
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2023, 7 (POPL): : 91 - 120
  • [45] LEARNING A NEAR-OPTIMAL ESTIMATOR FOR SURFACE SHAPE FROM SHADING
    KNILL, DC
    KERSTEN, D
    COMPUTER VISION GRAPHICS AND IMAGE PROCESSING, 1990, 50 (01): : 75 - 100
  • [46] Learning Near-Optimal Intrusion Responses Against Dynamic Attackers
    Hammar, Kim
    Stadler, Rolf
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2024, 21 (01): : 1158 - 1177
  • [47] Near-Optimal Φ-Regret Learning in Extensive-Form Games
    Anagnostides, Ioannis
    Farina, Gabriele
    Sandholm, Tuomas
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202 : 814 - 839
  • [48] Near-optimal Trajectory Tracking in Quadcopters using Reinforcement Learning
    Engelhardt, Randal
    Velazquez, Alberto
    Sardarmehni, Tohid
    IFAC PAPERSONLINE, 2024, 58 (28): : 61 - 65
  • [49] Near-Optimal Representation Learning for Linear Bandits and Linear RL
    Hu, Jiachen
    Chen, Xiaoyu
    Jin, Chi
    Li, Lihong
    Wang, Liwei
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [50] Polynomial-time reinforcement learning of near-optimal policies
    Pivazyan, K
    Shoham, Y
    EIGHTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-02)/FOURTEENTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-02), PROCEEDINGS, 2002, : 205 - 210