Towards provably efficient quantum algorithms for large-scale machine-learning models

被引:17
|
作者
Liu, Junyu [1 ,2 ,3 ,4 ,5 ,6 ]
Liu, Minzhao [7 ,8 ]
Liu, Jin-Peng [9 ,10 ,11 ]
Ye, Ziyu [2 ]
Wang, Yunfei [12 ]
Alexeev, Yuri [2 ,3 ,8 ]
Eisert, Jens [13 ]
Jiang, Liang [1 ,3 ]
机构
[1] Univ Chicago, Pritzker Sch Mol Engn, Chicago, IL 60637 USA
[2] Univ Chicago, Dept Comp Sci, Chicago, IL 60637 USA
[3] Chicago Quantum Exchange, Chicago, IL 60637 USA
[4] Univ Chicago, Kadanoff Ctr Theoret Phys, Chicago, IL 60637 USA
[5] qBraid Co, Chicago, IL 60615 USA
[6] SeQure, Chicago, IL 60615 USA
[7] Univ Chicago, Dept Phys, Chicago, IL 60637 USA
[8] Argonne Natl Lab, Computat Sci Div, Lemont, IL 60439 USA
[9] Univ Calif Berkeley, Simons Inst Theory Comp, Berkeley, CA 94720 USA
[10] Univ Calif Berkeley, Dept Math, Berkeley, CA 94720 USA
[11] MIT, Ctr Theoret Phys, Cambridge, MA 02139 USA
[12] Brandeis Univ, Martin A Fisher Sch Phys, Waltham, MA 02453 USA
[13] Free Univ Berlin, Dahlem Ctr Complex Quantum Syst, D-14195 Berlin, Germany
关键词
D O I
10.1038/s41467-023-43957-x
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Large machine learning models are revolutionary technologies of artificial intelligence whose bottlenecks include huge computational expenses, power, and time used both in the pre-training and fine-tuning process. In this work, we show that fault-tolerant quantum computing could possibly provide provably efficient resolutions for generic (stochastic) gradient descent algorithms, scaling as O(T-2 x polylog(n)), where n is the size of the models and T is the number of iterations in the training, as long as the models are both sufficiently dissipative and sparse, with small learning rates. Based on earlier efficient quantum algorithms for dissipative differential equations, we find and prove that similar algorithms work for (stochastic) gradient descent, the primary algorithm for machine learning. In practice, we benchmark instances of large machine learning models from 7 million to 103 million parameters. We find that, in the context of sparse training, a quantum enhancement is possible at the early stage of learning after model pruning, motivating a sparse parameter download and re-upload scheme. Our work shows solidly that fault-tolerant quantum algorithms could potentially contribute to most state-of-the-art, large-scale machine-learning problems.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Efficient algorithms for computing the best subset regression models for large-scale problems
    Hofmann, Marc
    Gatu, Cristian
    Kontoghiorghes, Erricos John
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2007, 52 (01) : 16 - 29
  • [32] Efficient Distributed Transfer Learning for Large-Scale Gaussian Graphic Models
    Zhou, Xingcai
    Zheng, Haotian
    Zhang, Haoran
    Huang, Chao
    STAT, 2024, 13 (04):
  • [33] Efficient Large-Scale Structured Learning
    Branson, Steve
    Beijbom, Oscar
    Belongie, Serge
    2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 1806 - 1813
  • [34] StableToolBench: Towards Stable Large-Scale Benchmarking on Tool Learning of Large Language Models
    Guo, Zhicheng
    Cheng, Sijie
    Wang, Hao
    Liang, Shihao
    Qin, Yujia
    Li, Peng
    Liu, Zhiyuan
    Sun, Maosong
    Liu, Yang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 11143 - 11156
  • [35] Three-dimensional realizations of flood flow in large-scale rivers using the neural fuzzy-based machine-learning algorithms
    Zhang, Zexia
    Limaye, Ajay B.
    Khosronejad, Ali
    COMPUTERS & FLUIDS, 2022, 246
  • [36] Large-scale dynamic flood monitoring in an arid-zone floodplain using SAR data and hybrid machine-learning models
    Panahi, Mahdi
    Rahmati, Omid
    Kalantari, Zahra
    Darabi, Hamid
    Rezaie, Fatemeh
    Moghaddam, Davoud Davoudi
    Ferreira, Carla Sofia Santos
    Foody, Giles
    Aliramaee, Ramyar
    Bateni, Sayed M.
    Lee, Chang-Wook
    Lee, Saro
    JOURNAL OF HYDROLOGY, 2022, 611
  • [37] Randomized algorithms for large-scale dictionary learning
    Wu, Gang
    Yang, Jiali
    NEURAL NETWORKS, 2024, 179
  • [38] A Machine-Learning Pipeline for Large-Scale Power-Quality Forecasting in the Mexican Distribution Grid
    Flores, Juan J.
    Garcia-Nava, Jose L.
    Gonzalez, Jose R. Cedeno
    Tellez, Victor M.
    Calderon, Felix
    Medrano, Arturo
    APPLIED SCIENCES-BASEL, 2022, 12 (17):
  • [39] Scalable Algorithms for Bayesian Inference of Large-Scale Models from Large-Scale Data
    Ghattas, Omar
    Isaac, Tobin
    Petra, Noemi
    Stadler, Georg
    HIGH PERFORMANCE COMPUTING FOR COMPUTATIONAL SCIENCE - VECPAR 2016, 2017, 10150 : 3 - 6
  • [40] Towards Scalable Querying of Large-Scale Models
    Barmpis, Konstantinos
    Kolovos, Dimitrios S.
    MODELLING FOUNDATIONS AND APPLICATIONS, ECMFA 2014, 2014, 8569 : 35 - 50