Unsupervised Pre-training on Improving the Performance of Neural Network in Regression

被引:0
|
作者
Salida, Pallabi [1 ]
Vij, Prateek [1 ]
Baruah, Rashmi Dutta [1 ]
机构
[1] Indian Inst Technol Guwahati, Dept Comp Sci & Engn, Gauhati 781039, Assam, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The paper aims to empirically analyse the performance of the prediction capability of Artificial Neural Network by applying a pre-training mechanism. The pre-training used here is same as the training of Deep Belief Network where the network is formed by stacking Restricted Boltzmann Machine one above the other successively. A different set of experiments are performed to understand in what scenario pre-trained ANN performed better than randomly initialised ANN. The results of experiments showed that pre-trained model performed better than randomly initialised ANN in terms of generalised error, computational units required and most importantly robust to change in hyperparameters such as learning rate and model architecture. The only cost is in additional time involved in the pre-training phase. Further, the learned knowledge in pre-training, which is stored as weights in ANN, are analysed using Hinton diagram. The analysis could provide clear picture of the pre-training that learned some of the hidden characteristics of the data.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] UNSUPERVISED POINT CLOUD PRE-TRAINING VIA CONTRASTING AND CLUSTERING
    Mei, Guofeng
    Huang, Xiaoshui
    Liu, Juan
    Zhang, Jian
    Wu, Qiang
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 66 - 70
  • [42] Neural Graph Matching for Pre-training Graph Neural Networks
    Hou, Yupeng
    Hu, Binbin
    Zhao, Wayne Xin
    Zhang, Zhiqiang
    Zhou, Jun
    Wen, Ji-Rong
    PROCEEDINGS OF THE 2022 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2022, : 172 - 180
  • [43] Improving the Sample Efficiency of Pre-training Language Models
    Berend, Gabor
    ERCIM NEWS, 2024, (136): : 38 - 40
  • [44] Improving Monocular Depth Estimation by Semantic Pre-training
    Rottmann, Peter
    Posewsky, Thorbjorn
    Milioto, Andres
    Stachniss, Cyrill
    Behley, Jens
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 5916 - 5923
  • [45] SpanBERT: Improving Pre-training by Representing and Predicting Spans
    Joshi, Mandar
    Chen, Danqi
    Liu, Yinhan
    Weld, Daniel S.
    Zettlemoyer, Luke
    Levy, Omer
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 : 64 - 77
  • [46] Improving Reinforcement Learning Pre-Training with Variational Dropout
    Blau, Tom
    Ott, Lionel
    Ramos, Fabio
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 4115 - 4122
  • [47] Layer-wise Pre-training Mechanism Based on Neural Network for Epilepsy Detection
    Lin, Zichao
    Gu, Zhenghui
    Li, Yinghao
    Yu, Zhuliang
    Li, Yuanqing
    2020 12TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2020, : 224 - 227
  • [48] Selection of an artificial pre-training neural network for the classification of inland vessels based on their images
    Bobkowska, Katarzyna
    Bodus-Olkowska, Izabela
    SCIENTIFIC JOURNALS OF THE MARITIME UNIVERSITY OF SZCZECIN-ZESZYTY NAUKOWE AKADEMII MORSKIEJ W SZCZECINIE, 2021, 67 (139):
  • [49] A deep neural network model for Chinese toponym matching with geographic pre-training model
    Qiu, Qinjun
    Zheng, Shiyu
    Tian, Miao
    Li, Jiali
    Ma, Kai
    Tao, Liufeng
    Xie, Zhong
    INTERNATIONAL JOURNAL OF DIGITAL EARTH, 2024, 17 (01)
  • [50] LORE plus plus : Logical location regression network for table structure recognition with pre-training
    Long, Rujiao
    Xing, Hangdi
    Yang, Zhibo
    Zheng, Qi
    Yu, Zhi
    Huang, Fei
    PATTERN RECOGNITION, 2024, 157