Language Modeling with a General Second-Order RNN

被引:0
|
作者
Maupome, Diego [1 ]
Meurs, Marie-Jean [1 ]
机构
[1] Univ Quebec Montreal UQAM, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会; 加拿大健康研究院;
关键词
Recurrent Neural Networks; Language Modeling;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Different Recurrent Neural Network (RNN) architectures update their state in different manners as the input sequence is processed. RNNs including a multiplicative interaction between their current state and the current input, second-order ones, show promising performance in language modeling. In this paper, we introduce a second-order RNNs that generalizes existing ones. Evaluating on the Penn Treebank dataset, we analyze how its different components affect its performance in character-lever recurrent language modeling. We perform our experiments controlling the parameter counts of models. We find that removing the first-order terms does not hinder performance. We perform further experiments comparing the effects of the relative size of the state space and the multiplicative interaction space on performance. Our expectation was that a larger states would benefit language models built on longer documents, and larger multiplicative interaction states would benefit ones built on larger input spaces. However, our results suggest that this is not the case and the optimal relative size is the same for both document tokenizations used.
引用
收藏
页码:4749 / 4753
页数:5
相关论文
共 50 条