Modeling language and cognition with deep unsupervised learning: a tutorial overview

被引:46
|
作者
Zorzi, Marco [1 ,2 ]
Testolin, Alberto [1 ]
Stoianov, Ivilin P. [1 ,3 ]
机构
[1] Univ Padua, Dept Gen Psychol, Computat Cognit Neurosci Lab, I-35131 Padua, Italy
[2] IRCCS San Camillo Neurorehabil Hosp, Venice, Italy
[3] CNR, Inst Cognit Sci & Technol, Rome, Italy
来源
FRONTIERS IN PSYCHOLOGY | 2013年 / 4卷
关键词
neural networks; connectionist modeling; deep learning; hierarchical generative models; unsupervised learning; visual word recognition; PROBABILISTIC MODELS; CONNECTIONIST; REPRESENTATIONS; ORGANIZATION; RECOGNITION; PRINCIPLES; EMERGENCE; ALGORITHM; DYSLEXIA; WORDS;
D O I
10.3389/fpsyg.2013.00515
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Deep unsupervised learning in stochastic recurrent neural networks with many layers of hidden units is a recent breakthrough in neural computation research. These networks build a hierarchy of progressively more complex distributed representations of the sensory data by fitting a hierarchical generative model. In this article we discuss the theoretical foundations of this approach and we review key issues related to training, testing and analysis of deep networks for modeling language and cognitive processing. The classic letter and word perception problem of McClelland and Rumelhart (1981) is used as a tutorial example to illustrate how structured and abstract representations may emerge from deep generative learning. We argue that the focus on deep architectures and generative (rather than discriminative) learning represents a crucial step forward for the connectionist modeling enterprise, because it offers a more plausible model of cortical learning as well as a way to bridge the gap between emergentist connectionist models and structured Bayesian models of cognition.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Unsupervised Clustering for Deep Learning: A tutorial survey
    Karoly, Artur Istvan
    Fuller, Robert
    Galambos, Peter
    ACTA POLYTECHNICA HUNGARICA, 2018, 15 (08) : 29 - 53
  • [2] Unsupervised learning of dependency structure for language modeling
    Gao, JF
    Suzuki, H
    41ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2003, : 521 - 528
  • [3] Unsupervised learning of dependency structure for language modeling
    Gao, Jianfeng
    Suzuki, Hisami
    Proc. Annu. Meet. Assoc. Comput Linguist., 1600,
  • [4] Machine Learning and Deep Learning Techniques for Colocated MIMO Radars: A Tutorial Overview
    Davoli, Alessandro
    Guerzoni, Giorgio
    Vitetta, Giorgio M.
    IEEE ACCESS, 2021, 9 : 33704 - 33755
  • [5] Language/Dialect Recognition Based on Unsupervised Deep Learning
    Zhang, Qian
    Hansen, John H. L.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2018, 26 (05) : 873 - 882
  • [6] Fast and Accurate Deep Bidirectional Language Representations for Unsupervised Learning
    Shin, Joongbo
    Lee, Yoonhyung
    Yoon, Seunghyun
    Jung, Kyomin
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 823 - 835
  • [7] The tutorial for learning programming language
    Borza, S.
    Simion, C.
    3rd Balkan Region Conference on Engineering Education, Conference Proceedings: ADVANCING ENGINEERING EDUCATION, 2005, : 226 - 229
  • [8] Language breakdown and linguistic theory:: A tutorial overview
    Levy, Y
    Kavé, G
    LINGUA, 1999, 107 (1-2) : 95 - 143
  • [9] Language-Independent Text Tokenization Using Unsupervised Deep Learning
    Mahmoud, Hanan A. Hosni
    Hafez, Alaaeldin M.
    Alabdulkreem, Eatedal
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 35 (01): : 321 - 334
  • [10] A review of unsupervised feature learning and deep learning for time-series modeling
    Langkvist, Martin
    Karlsson, Lars
    Loutfi, Amy
    PATTERN RECOGNITION LETTERS, 2014, 42 : 11 - 24