Distributed Submodular Maximization

被引:0
|
作者
Mirzasoleiman, Baharan [1 ]
Karbasi, Amin [2 ]
Sarkar, Rik [3 ]
Krause, Andreas [1 ]
机构
[1] ETH, Dept Comp Sci, Univ Str 6, CH-8092 Zurich, Switzerland
[2] Yale Univ, Sch Engn & Appl Sci, New Haven, CT USA
[3] Univ Edinburgh, Dept Informat, 10 Crichton St, Edinburgh EH8 9AB, Midlothian, Scotland
关键词
distributed computing; submodular functions; approximation algorithms; greedy algorithms; map-reduce; SET; APPROXIMATIONS; ALGORITHM; NETWORKS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Many large-scale machine learning problems-clustering, non-parametric learning, kernel machines, etc.-require selecting a small yet representative subset from a large dataset. Such problems can often be reduced to maximizing a submodular set function subject to various constraints. Classical approaches to submodular optimization require centralized access to the full dataset, which is impractical for truly large-scale problems. In this paper, we consider the problem of submodular function maximization in a distributed fashion. We develop a simple, two-stage protocol GREEDI, that is easily implemented using MapReduce style computations. We theoretically analyze our approach, and show that under certain natural conditions, performance close to the centralized approach can be achieved. We begin with monotone submodular maximization subject to a cardinality constraint, and then extend this approach to obtain approximation guarantees for (not necessarily monotone) submodular maximization subject to more general constraints including matroid or knapsack constraints. In our extensive experiments, we demonstrate the effectiveness of our approach on several applications, including sparse Gaussian process inference and exemplar based clustering on tens of millions of examples using Hadoop.
引用
收藏
页数:44
相关论文
共 50 条
  • [31] Gradient Methods for Submodular Maximization
    Hassani, Hamed
    Soltanolkotabi, Mahdi
    Karbasi, Amin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [32] Practical Budgeted Submodular Maximization
    Moran Feldman
    Zeev Nutov
    Elad Shoham
    Algorithmica, 2023, 85 : 1332 - 1371
  • [33] The Power of Subsampling in Submodular Maximization
    Harshaw, Christopher
    Kazemi, Ehsan
    Feldman, Moran
    Karbasi, Amin
    MATHEMATICS OF OPERATIONS RESEARCH, 2022, 47 (02) : 1365 - 1393
  • [34] Improved Inapproximability for Submodular Maximization
    Austrin, Per
    APPROXIMATION, RANDOMIZATION, AND COMBINATORIAL OPTIMIZATION: ALGORITHMS AND TECHNIQUES, 2010, 6302 : 12 - 24
  • [35] The FAST Algorithm for Submodular Maximization
    Breuer, Adam
    Balkanski, Eric
    Singer, Yaron
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [36] Online Submodular Maximization with Preemption
    Buchbinder, Niv
    Feldman, Moran
    Schwartz, Roy
    ACM TRANSACTIONS ON ALGORITHMS, 2019, 15 (03)
  • [37] Horizontally Scalable Submodular Maximization
    Lucic, Mario
    Bachem, Olivier
    Zadimoghaddam, Morteza
    Krause, Andreas
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [38] Streaming adaptive submodular maximization*
    Tang, Shaojie
    Yuan, Jing
    THEORETICAL COMPUTER SCIENCE, 2023, 944
  • [39] On the Complexity of Dynamic Submodular Maximization
    Chen, Xi
    Peng, Binghui
    PROCEEDINGS OF THE 54TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING (STOC '22), 2022, : 1685 - 1698
  • [40] The FAST Algorithm for Submodular Maximization
    Breuer, Adam
    Balkanski, Eric
    Singer, Yaron
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,