Advances and Trends in Extreme Learning Machine

被引:0
|
作者
Xu R. [1 ]
Liang X. [1 ]
Qi J.-S. [1 ]
Li Z.-Y. [1 ]
Zhang S.-S. [1 ]
机构
[1] School of Information, Renmin University of China, Beijing
来源
基金
中国国家自然科学基金;
关键词
Deep learning; Extreme learning machine; Kernel learning; Network structure; Online sequential learning; Parallel computing; Regularization;
D O I
10.11897/SP.J.1016.2019.01640
中图分类号
学科分类号
摘要
Extreme Learning Machine (ELM) as a new single hidden layer feedforward neural network (SLFN) learning framework has obtained extensive attention and in-depth research in various domains. It has been widely used in many applications, such as action recognition, emotion recognition, fault diagnosis, and so on. ELM was originally proposed for "generalized" single hidden layer feedforward neural networks to overcome the challenging issues faced by back-propagation (BP) learning algorithm and its variants. Recent studies show that ELM can be extended to "generalized" multilayer feedforward neural networks in which a hidden node could be a subnetwork of nodes or a combination of other hidden nodes. ELM provides an efficient and unified learning framework for regression, classification, feature learning, and clustering. The learning theories of ELM show that when learning parameters of hidden layer nodes are generated independently of training samples, as long as the activation function of feedforward neural network is non-linear and continuous, it can approach any continuous objective function or any complex decision boundary in the classification task. In ELM, the input weights and hidden biases connecting the input layer and the hidden layer can be independent of the training sample and randomly generated from any continuous probability distribution. The output weight matrix between the hidden layer and the output layer is obtained by minimizing the square loss function and solving the Moore-Penrose generalized inverse operation to obtain the minimum norm least squares solution. The only parameter that needs to be optimized is the number of hidden layer nodes. It has been shown by theoretical studies that ELM is capable of maintaining the universal approximation and classification capability of SLFNs even if it works with randomly generated hidden nodes. Different from traditional gradient-based neural network learning algorithms, which are sensitive to the combination of parameters and easy to trap in local optimum, ELM has faster learning speed, least human intervention and easy to implementation. In a word, ELM has become one of the most popular research directions in the field of artificial intelligence in recent years and received widespread attention from more and more research members domestic and abroad. To make it more suitable and efficient for specific applications, ELM theories and algorithms have been investigated extensively in the past few decades. Recently, random neurons have gradually been used in deep learning and ELM provides the theoretical basis for use. This paper aims to provide a comprehensive review of existing research results in ELM. We first give an introduction to the historical background and developments of ELM. Then we describe the principle and algorithm of ELM in detail followed by the introduction of its feature map and feature space. After an overview of ELM theory, we discuss and analyze state-of-the-art algorithms or the typical variants of ELM, including models, solution approaches and relevant problems. On this basis, the core ideas, advantages and disadvantages of each algorithm are summarized. Furthermore, the latest applications of ELM are reviewed. In the end, several controversies, open issues and challenges in ELM are pointed out together with its future research directions and trends. © 2019, Science Press. All right reserved.
引用
收藏
页码:1640 / 1670
页数:30
相关论文
共 150 条
  • [1] Huang G.B., Zhu Q.Y., Siew C.K., Extreme learning machine: A new learning scheme of feedforward neural networks, Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, pp. 985-990, (2004)
  • [2] Chyzhyk D., Savio A., Grana M., Computer aided diagnosis of schizophrenia on resting state fMRI data by ensembles of ELM, Neural Networks, 68, pp. 23-33, (2015)
  • [3] Huang Z., Yu Y., Gu J., An efficient method for traffic sign recognition based on extreme learning machine, IEEE Transactions on Cybernetics, 47, 4, pp. 920-933, (2017)
  • [4] Wang S., Deng C., Lin W., NMF-based image quality assessment using extreme learning machine, IEEE Transactions on Cybernetics, 47, 1, pp. 232-243, (2017)
  • [5] Huang G.B., Siew C.K., Extreme learning machine: RBF network case, Proceedings of the 8th International Conference on Control, Automation, Robotics and Vision(ICARCV), pp. 1029-1036, (2004)
  • [6] Zhang Y., Wu J., Zhou C., Instance cloned extreme learning machine, Pattern Recognition, 68, pp. 52-65, (2017)
  • [7] Ertugrul O.F., Forecasting electricity load by a novel recurrent extreme learning machines approach, International Journal of Electrical Power & Energy Systems, 78, pp. 429-435, (2016)
  • [8] Tang J., Deng C., Huang G.B., Extreme learning machine for multilayer perceptron, IEEE Transactions on Neural Networks and Learning Systems, 27, 4, pp. 809-821, (2016)
  • [9] Huang G.B., Zhu Q.Y., Siew C.K., Extreme learning machine: Theory and applications, Neurocomputing, 70, 1-3, pp. 489-501, (2006)
  • [10] Huang G.B., Chen L., Convex incremental extreme learning machine, Neurocomputing, 70, 16-18, pp. 3056-3062, (2007)