MLink: Linking Black-Box Models From Multiple Domains for Collaborative Inference

被引:1
|
作者
Yuan, Mu [1 ]
Zhang, Lan [1 ,2 ]
Zheng, Zimu [3 ]
Zhang, Yi-Nan [1 ]
Li, Xiang-Yang [1 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei 230026, Anhui, Peoples R China
[2] Hefei Comprehens Natl Sci Ctr, Inst Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[3] Huawei Cloud, Edge Cloud Innovat Lab, Shenzhen 518129, Guangdong, Peoples R China
关键词
Model linking; multi-model inference;
D O I
10.1109/TPAMI.2023.3283780
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The cost efficiency of model inference is critical to real-world machine learning (ML) applications, especially for delay-sensitive tasks and resource-limited devices. A typical dilemma is: in order to provide complex intelligent services (e.g., smart city), we need inference results of multiple ML models, but the cost budget (e.g., GPU memory) is not enough to run all of them. In this work, we study underlying relationships among black-box ML models and propose a novel learning task: model linking, which aims to bridge the knowledge of different black-box models by learning mappings (dubbed model links) between their output spaces. We propose the design of model links which supports linking heterogeneous black-box ML models. Also, in order to address the distribution discrepancy challenge, we present adaptation and aggregation methods of model links. Based on our proposed model links, we developed a scheduling algorithm, named MLink. Through collaborative multi-model inference enabled by model links, MLink can improve the accuracy of obtained inference results under the cost budget. We evaluated MLink on a multi-modal dataset with seven different ML models and two real-world video analytics systems with six ML models and 3,264 hours of video. Experimental results show that our proposed model links can be effectively built among various black-box models. Under the budget of GPU memory, MLink can save 66.7% inference computations while preserving 94% inference accuracy, which outperforms multi-task learning, deep reinforcement learning-based scheduler and frame filtering baselines.
引用
收藏
页码:12085 / 12097
页数:13
相关论文
共 50 条
  • [1] MLink: Linking Black-Box Models for Collaborative Multi-Model Inference
    Yuan, Mu
    Zhang, Lan
    Li, Xiang-Yang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 9475 - 9483
  • [2] Scalable Inference for Gaussian Process Models with Black-Box Likelihoods
    Dezfouli, Amir
    Bonilla, Edwin V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [3] Black-box Bayesian inference for agent-based models
    Dyer, Joel
    Cannon, Patrick
    Farmer, J. Doyne
    Schmon, Sebastian M.
    JOURNAL OF ECONOMIC DYNAMICS & CONTROL, 2024, 161
  • [4] On the Convergence of Black-Box Variational Inference
    Kim, Kyurae
    Oh, Jisu
    Wu, Kaiwen
    Ma, Yi-An
    Gardner, Jacob R.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Black-box Coreset Variational Inference
    Manousakas, Dionysis
    Ritter, Hippolyt
    Karaletsos, Theofanis
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [6] Approximate posterior inference for Bayesian models: black-box expectation propagation
    Ximing Li
    Changchun Li
    Jinjin Chi
    Jihong Ouyang
    Knowledge and Information Systems, 2022, 64 : 2361 - 2387
  • [7] Approximate posterior inference for Bayesian models: black-box expectation propagation
    Li, Ximing
    Li, Changchun
    Chi, Jinjin
    Ouyang, Jihong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2022, 64 (09) : 2361 - 2387
  • [8] Black-Box Inference for Non-Linear Latent Force Models
    Ward, Wil O. C.
    Ryder, Tom
    Prangle, Dennis
    Alvarez, Mauricio A.
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [9] InferDPT: Privacy-preserving Inference for Black-box Large Language Models
    Tong, Meng
    Chen, Kejiang
    Zhang, Jie
    Qi, Yuang
    Zhang, Weiming
    Yu, Nenghai
    Zhang, Tianwei
    Zhang, Zhikun
    arXiv, 2023,
  • [10] Black-Box Test Generation from Inferred Models
    Papadopoulos, Petros
    Walkinshaw, Neil
    2015 IEEE/ACM FOURTH INTERNATIONAL WORKSHOP ON REALIZING ARTIFICIAL INTELLIGENCE SYNERGIES IN SOFTWARE ENGINEERING (RAISE 2015), 2015, : 19 - 24