Post-training discriminative pruning for RBMs

被引:3
|
作者
Sanchez-Gutierrez, Maximo [1 ]
Albornoz, Enrique M. [2 ]
Rufiner, Hugo L. [2 ,3 ]
Goddard Close, John [1 ]
机构
[1] Univ Autonoma Metropolitana, Dept Ingn Elect, Iztapalapa, Mexico
[2] UNL, CONICET, Inst Invest Senales Sistemas & Inteligencia Compu, FICH,Sinc I, Ciudad Univ,S3000, Paraje El Pozo, Santa Fe, Argentina
[3] UNER, Fac Ingn, Lab Cibernet, Oro Verde, Entre Rios, Argentina
关键词
Restricted Boltzmann machines; Pruning; Discriminative information; Phoneme classification; Emotion classification; BOLTZMANN MACHINES; NEURAL-NETWORKS; DEEP; ALGORITHM; SPEECH;
D O I
10.1007/s00500-017-2784-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the major challenges in the area of artificial neural networks is the identification of a suitable architecture for a specific problem. Choosing an unsuitable topology can exponentially increase the training cost, and even hinder network convergence. On the other hand, recent research indicates that larger or deeper nets can map the problem features into a more appropriate space, and thereby improve the classification process, thus leading to an apparent dichotomy. In this regard, it is interesting to inquire whether independent measures, such as mutual information, could provide a clue to finding the most discriminative neurons in a network. In the present work, we explore this question in the context of Restricted Boltzmann machines, by employing different measures to realize post-training pruning. The neurons which are determined by each measure to be the most discriminative, are combined and a classifier is applied to the ensuing network to determine its usefulness. We find that two measures in particular seem to be good indicators of the most discriminative neurons, producing savings of generally more than 50% of the neurons, while maintaining an acceptable error rate. Further, it is borne out that starting with a larger network architecture and then pruning is more advantageous than using a smaller network to begin with. Finally, a quantitative index is introduced which can provide information on choosing a suitable pruned network.
引用
下载
收藏
页码:767 / 781
页数:15
相关论文
共 50 条
  • [31] Discriminative Pruning for Discriminative ITG Alignment
    Liu, Shujie
    Li, Chi-Ho
    Zhou, Ming
    ACL 2010: 48TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2010, : 316 - 324
  • [32] Pre-training and post-training carbohydrate consumption in squash players
    de Moura, Monica Goncalves
    Perea, Carolina
    Stulbach, Tamara
    Caparros, Daniele Ramos
    RBNE-REVISTA BRASILEIRA DE NUTRICAO ESPORTIVA, 2014, 8 (47): : 343 - 350
  • [33] Post-training meditation mediates motor memory consolidation
    Immink, Maarten A.
    JOURNAL OF SPORT & EXERCISE PSYCHOLOGY, 2012, 34 : S90 - S90
  • [34] Post-training massage: A review for strength and power athletes
    Chiu, LZ
    Weiss, LW
    Fry, AC
    STRENGTH AND CONDITIONING JOURNAL, 2001, 23 (04): : 65 - 69
  • [35] POST-TRAINING HYPOXIA ENHANCES AVOIDANCE CONDITIONING IN CHICK
    CHERKIN, A
    PHARMACOLOGIST, 1970, 12 (02): : 271 - &
  • [36] POST-TRAINING QUANTIZATION FOR VISION TRANSFORMER IN TRANSFORMED DOMAIN
    Feng, Kai
    Chen, Zhuo
    Gao, Fei
    Wang, Zhe
    Xu, Long
    Lin, Weisi
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1457 - 1462
  • [37] Post-training Quantization Methods for Deep Learning Models
    Kluska, Piotr
    Zieba, Maciej
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS (ACIIDS 2020), PT I, 2020, 12033 : 467 - 479
  • [38] PTQD: Accurate Post-Training Quantization for Diffusion Models
    He, Yefei
    Liu, Luping
    Liu, Jing
    Wu, Weijia
    Zhou, Hong
    Zhuang, Bohan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [39] EFFECT OF POST-TRAINING EXPOSURE TO AN AVERSIVE STIMULUS ON RETENTION
    WHITE, NM
    LEGREE, P
    PHYSIOLOGICAL PSYCHOLOGY, 1984, 12 (03): : 233 - 236
  • [40] Exploring the impact of post-training rounding in regression models
    Jan Kalina
    Applications of Mathematics, 2024, 69 : 257 - 271