Training Fully Connected Neural Networks is ∃R-Complete

被引:0
|
作者
Bertschinger, Daniel [1 ]
Hertrich, Christoph [2 ,5 ,6 ]
Jungeblut, Paul [3 ]
Miltzow, Tillmann [4 ]
Weber, Simon [1 ]
机构
[1] Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[2] London Sch Econ & Polit Sci, Dept Math, London, England
[3] Karlsruhe Inst Technol, Inst Theoret Informat, Karlsruhe, Germany
[4] Univ Utrecht, Dept Informat & Comp Sci, Utrecht, Netherlands
[5] Univ Libre, Brussels, Belgium
[6] Goethe Univ, Frankfurt, Germany
基金
欧洲研究理事会; 瑞士国家科学基金会;
关键词
BOUNDS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully connected neural network to fit a given set of data points, also known as empirical risk minimization. We show that the problem is there exists R-complete. This complexity class can be defined as the set of algorithmic problems that are polynomial-time equivalent to finding real roots of a multivariate polynomial with integer coefficients. Furthermore, we show that arbitrary algebraic numbers are required as weights to be able to train some instances to optimality, even if all data points are rational. Our result already applies to fully connected instances with two inputs, two outputs, and one hidden layer of ReLU neurons. Thereby, we strengthen a result by Abrahamsen, Kleist and Miltzow [NeurIPS 2021]. A consequence of this is that a combinatorial search algorithm like the one by Arora, Basu, Mianjy and Mukherjee [ICLR 2018] is impossible for networks with more than one output dimension, unless NP = there exists R.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Fully forward mode training for optical neural networks
    Xue, Zhiwei
    Zhou, Tiankuang
    Xu, Zhihao
    Yu, Shaoliang
    Dai, Qionghai
    Fang, Lu
    NATURE, 2024, 632 (8024) : 280 - 286
  • [32] TRAINING FULLY RECURRENT NEURAL NETWORKS WITH COMPLEX WEIGHTS
    KECHRIOTIS, G
    MANOLAKOS, ES
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-ANALOG AND DIGITAL SIGNAL PROCESSING, 1994, 41 (03): : 235 - 238
  • [33] Structural Damage Identification of Composite Rotors Based on Fully Connected Neural Networks and Convolutional Neural Networks
    Scholz, Veronika
    Winkler, Peter
    Hornig, Andreas
    Gude, Maik
    Filippatos, Angelos
    SENSORS, 2021, 21 (06) : 1 - 15
  • [34] Training fully connected networks with resistive memories: impact of device failures
    Romero, Louis P.
    Ambrogio, Stefano
    Giordano, Massimo
    Cristiano, Giorgio
    Bodini, Martina
    Narayanan, Pritish
    Tsai, Hsinyu
    Shelby, Robert M.
    Burr, Geoffrey W.
    FARADAY DISCUSSIONS, 2019, 213 : 371 - 391
  • [35] Estimation of a regression function on a manifold by fully connected deep neural networks
    Kohler, Michael
    Langer, Sophie
    Reif, Ulrich
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2023, 222 : 160 - 181
  • [36] Modeling Dynamic Hysteresis through Fully Connected Cascade Neural Networks
    Laudani, Antonino
    Lozito, Gabriele Maria
    Fulginei, Francesco Riganti
    Salvini, Alessandro
    2016 IEEE 2ND INTERNATIONAL FORUM ON RESEARCH AND TECHNOLOGIES FOR SOCIETY AND INDUSTRY LEVERAGING A BETTER TOMORROW (RTSI), 2016, : 387 - 391
  • [37] Generalization in fully-connected neural networks for time series forecasting
    Borovykh, Anastasia
    Oosterlee, Cornelis W.
    Bohte, Sander M.
    JOURNAL OF COMPUTATIONAL SCIENCE, 2019, 36
  • [38] AirFC: Designing Fully Connected Layers for Neural Networks with Wireless Signals
    Reus-Muns, Guillem
    Alemdar, Kubra
    Sanchez, Sara Garcia
    Roy, Debashri
    Chowdhury, Kaushik R.
    PROCEEDINGS OF THE 2023 INTERNATIONAL SYMPOSIUM ON THEORY, ALGORITHMIC FOUNDATIONS, AND PROTOCOL DESIGN FOR MOBILE NETWORKS AND MOBILE COMPUTING, MOBIHOC 2023, 2023, : 71 - 80
  • [39] THERMODYNAMIC PROPERTIES OF FULLY CONNECTED Q-ISING NEURAL NETWORKS
    BOLLE, D
    RIEGER, H
    SHIM, GM
    JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1994, 27 (10): : 3411 - 3426
  • [40] Fully Connected Feedforward Neural Networks Based CSI Feedback Algorithm
    Ming Gao
    Tanming Liao
    Yubin Lu
    中国通信, 2021, 18 (01) : 43 - 48