Modeling language and cognition with deep unsupervised learning: a tutorial overview

被引:46
|
作者
Zorzi, Marco [1 ,2 ]
Testolin, Alberto [1 ]
Stoianov, Ivilin P. [1 ,3 ]
机构
[1] Univ Padua, Dept Gen Psychol, Computat Cognit Neurosci Lab, I-35131 Padua, Italy
[2] IRCCS San Camillo Neurorehabil Hosp, Venice, Italy
[3] CNR, Inst Cognit Sci & Technol, Rome, Italy
来源
FRONTIERS IN PSYCHOLOGY | 2013年 / 4卷
关键词
neural networks; connectionist modeling; deep learning; hierarchical generative models; unsupervised learning; visual word recognition; PROBABILISTIC MODELS; CONNECTIONIST; REPRESENTATIONS; ORGANIZATION; RECOGNITION; PRINCIPLES; EMERGENCE; ALGORITHM; DYSLEXIA; WORDS;
D O I
10.3389/fpsyg.2013.00515
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Deep unsupervised learning in stochastic recurrent neural networks with many layers of hidden units is a recent breakthrough in neural computation research. These networks build a hierarchy of progressively more complex distributed representations of the sensory data by fitting a hierarchical generative model. In this article we discuss the theoretical foundations of this approach and we review key issues related to training, testing and analysis of deep networks for modeling language and cognitive processing. The classic letter and word perception problem of McClelland and Rumelhart (1981) is used as a tutorial example to illustrate how structured and abstract representations may emerge from deep generative learning. We argue that the focus on deep architectures and generative (rather than discriminative) learning represents a crucial step forward for the connectionist modeling enterprise, because it offers a more plausible model of cortical learning as well as a way to bridge the gap between emergentist connectionist models and structured Bayesian models of cognition.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] IPDPS 2017 NVIDIA Deep Learning Tutorial
    Bernauer, Julie
    2017 31ST IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2017, : 491 - 491
  • [42] Tutorial on Deep Learning Interpretation: A Data Perspective
    Yang, Zhou
    Liu, Ninghao
    Hu, Xia Ben
    Jin, Fang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 5156 - 5159
  • [43] Modeling Hebb Learning Rule for Unsupervised Learning
    Liu, Jia
    Gong, Maoguo
    Miao, Qiguang
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2315 - 2321
  • [44] An Overview of Unsupervised Deep Feature Representation for Text Categorization
    Wang, Shiping
    Cai, Jinyu
    Lin, Qihao
    Guo, Wenzhong
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2019, 6 (03) : 504 - 517
  • [45] Improve Deep Learning with Unsupervised Objective
    Zhang, Shufei
    Huang, Kaizhu
    Zhang, Rui
    Hussain, Amir
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT I, 2017, 10634 : 720 - 728
  • [46] An Unsupervised Deep Learning Framework via Integrated Optimization of Representation Learning and GMM-Based Modeling
    Wang, Jinghua
    Jiang, Jianmin
    COMPUTER VISION - ACCV 2018, PT I, 2019, 11361 : 249 - 265
  • [47] OMG Systems Modeling Language (OMG SysML™) Tutorial
    Friedenthal, Sanford
    Moore, Alan
    Steiner, Rick
    INCOSE Int. Sym., 2008, 1 (1731-1862): : 1731 - 1862
  • [48] Deep Bayesian Unsupervised Lifelong Learning
    Zhao, Tingting
    Wang, Zifeng
    Masoomi, Aria
    Dy, Jennifer
    NEURAL NETWORKS, 2022, 149 : 95 - 106
  • [49] Unsupervised Deep Learning by Neighbourhood Discovery
    Huang, Jiabo
    Dong, Qi
    Gong, Shaogang
    Zhu, Xiatian
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [50] Deep unsupervised learning of visual similarities
    Sanakoyeu, Artsiom
    Bautista, Miguel A.
    Ommer, Bjoern
    PATTERN RECOGNITION, 2018, 78 : 331 - 343