Unsupervised Pre-training on Improving the Performance of Neural Network in Regression

被引:0
|
作者
Salida, Pallabi [1 ]
Vij, Prateek [1 ]
Baruah, Rashmi Dutta [1 ]
机构
[1] Indian Inst Technol Guwahati, Dept Comp Sci & Engn, Gauhati 781039, Assam, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The paper aims to empirically analyse the performance of the prediction capability of Artificial Neural Network by applying a pre-training mechanism. The pre-training used here is same as the training of Deep Belief Network where the network is formed by stacking Restricted Boltzmann Machine one above the other successively. A different set of experiments are performed to understand in what scenario pre-trained ANN performed better than randomly initialised ANN. The results of experiments showed that pre-trained model performed better than randomly initialised ANN in terms of generalised error, computational units required and most importantly robust to change in hyperparameters such as learning rate and model architecture. The only cost is in additional time involved in the pre-training phase. Further, the learned knowledge in pre-training, which is stored as weights in ANN, are analysed using Hinton diagram. The analysis could provide clear picture of the pre-training that learned some of the hidden characteristics of the data.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Unsupervised Extractive Summarization by Pre-training Hierarchical Transformers
    Xu, Shusheng
    Zhang, Xingxing
    Wu, Yi
    Wei, Furu
    Zhou, Ming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1784 - 1795
  • [22] Investigating of Disease Name Normalization Using Neural Network and Pre-Training
    Lou, Yinxia
    Qian, Tao
    Li, Fei
    Zhou, Junxiang
    Ji, Donghong
    Cheng, Ming
    IEEE ACCESS, 2020, 8 : 85729 - 85739
  • [23] Unsupervised Pre-training for Temporal Action Localization Tasks
    Zhang, Can
    Yang, Tianyu
    Weng, Junwu
    Cao, Meng
    Wang, Jue
    Zou, Yuexian
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 14011 - 14021
  • [24] Exploring unsupervised pre-training for echo state networks
    Steiner, Peter
    Jalalvand, Azarakhsh
    Birkholz, Peter
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (34): : 24225 - 24242
  • [25] Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation
    Skurzhanskyi, O. H.
    Marchenko, O. O.
    Anisimov, A. V.
    CYBERNETICS AND SYSTEMS ANALYSIS, 2024, 60 (02) : 167 - 174
  • [26] Specialized Pre-Training of Neural Networks on Synthetic Data for Improving Paraphrase Generation
    O. H. Skurzhanskyi
    O. O. Marchenko
    A. V. Anisimov
    Cybernetics and Systems Analysis, 2024, 60 : 167 - 174
  • [27] A Task-Generic High-Performance Unsupervised Pre-Training Framework for ECG
    Wang, Guoxin
    Wang, Qingyuan
    Nag, Avishek
    John, Deepu
    IEEE SENSORS JOURNAL, 2024, 24 (13) : 21776 - 21785
  • [28] PERFORMANCE-EFFICIENCY TRADE-OFFS IN UNSUPERVISED PRE-TRAINING FOR SPEECH RECOGNITION
    Wu, Felix
    Kim, Kwangyoun
    Pan, Jing
    Han, Kyu J.
    Weinberger, Kilian Q.
    Artzi, Yoav
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7667 - 7671
  • [29] Heterogeneous graph convolutional network pre-training as side information for improving recommendation
    Do, Phuc
    Pham, Phu
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (18): : 15945 - 15961
  • [30] Heterogeneous graph convolutional network pre-training as side information for improving recommendation
    Phuc Do
    Phu Pham
    Neural Computing and Applications, 2022, 34 : 15945 - 15961