Unsupervised Pre-training on Improving the Performance of Neural Network in Regression

被引:0
|
作者
Salida, Pallabi [1 ]
Vij, Prateek [1 ]
Baruah, Rashmi Dutta [1 ]
机构
[1] Indian Inst Technol Guwahati, Dept Comp Sci & Engn, Gauhati 781039, Assam, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The paper aims to empirically analyse the performance of the prediction capability of Artificial Neural Network by applying a pre-training mechanism. The pre-training used here is same as the training of Deep Belief Network where the network is formed by stacking Restricted Boltzmann Machine one above the other successively. A different set of experiments are performed to understand in what scenario pre-trained ANN performed better than randomly initialised ANN. The results of experiments showed that pre-trained model performed better than randomly initialised ANN in terms of generalised error, computational units required and most importantly robust to change in hyperparameters such as learning rate and model architecture. The only cost is in additional time involved in the pre-training phase. Further, the learned knowledge in pre-training, which is stored as weights in ANN, are analysed using Hinton diagram. The analysis could provide clear picture of the pre-training that learned some of the hidden characteristics of the data.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Neural speech enhancement with unsupervised pre-training and mixture training
    Hao, Xiang
    Xu, Chenglin
    Xie, Lei
    NEURAL NETWORKS, 2023, 158 : 216 - 227
  • [2] Unsupervised Pre-training for Fully Convolutional Neural Networks
    Wiehman, Stiaan
    Kroon, Steve
    de Villiers, Hendrik
    2016 PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA AND ROBOTICS AND MECHATRONICS INTERNATIONAL CONFERENCE (PRASA-ROBMECH), 2016,
  • [3] An Empirical Study on Unsupervised Pre-training Approaches in Regression Problems
    Saikia, Pallabi
    Baruah, Rashmi Dutta
    2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 342 - 349
  • [4] Autoencoder Feature Residuals for Network Intrusion Detection: Unsupervised Pre-training for Improved Performance
    Lewandowski, Brian
    Paffenroth, Randy
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 1334 - 1341
  • [5] MALUP: A Malware Classification Framework using Convolutional Neural Network with Deep Unsupervised Pre-training
    Qiang, Qian
    Cheng, Mian
    Zhou, Yuan
    Ding, Yu
    Qi, Zisen
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 627 - 634
  • [6] Unsupervised Pre-Training for Detection Transformers
    Dai, Zhigang
    Cai, Bolun
    Lin, Yugeng
    Chen, Junying
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 12772 - 12782
  • [7] Unsupervised Pre-Training for Voice Activation
    Kolesau, Aliaksei
    Sesok, Dmitrij
    APPLIED SCIENCES-BASEL, 2020, 10 (23): : 1 - 13
  • [8] Improving Fractal Pre-training
    Anderson, Connor
    Farrell, Ryan
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 2412 - 2421
  • [9] Neural Grammatical Error Correction Systems with Unsupervised Pre-training on Synthetic Data
    Grundkiewicz, Roman
    Junczys-Dowmunt, Marcin
    Heafield, Kenneth
    INNOVATIVE USE OF NLP FOR BUILDING EDUCATIONAL APPLICATIONS, 2019, : 252 - 263
  • [10] Unsupervised Pre-Training with Spiking Neural Networks in Semi-Supervised Learning
    Dorogyy, Yaroslav
    Kolisnichenko, Vadym
    2018 IEEE FIRST INTERNATIONAL CONFERENCE ON SYSTEM ANALYSIS & INTELLIGENT COMPUTING (SAIC), 2018, : 177 - 180