Information encoding by deep neural networks: what can we learn?

被引:3
|
作者
ten Bosch, L. [1 ,2 ]
Boves, L. [1 ]
机构
[1] Radboud Univ Nijmegen, Nijmegen, Netherlands
[2] Max Planck Inst Psycholinguist, Nijmegen, Netherlands
关键词
deep neural networks; conventional knowledge; information encoding; structure discovery;
D O I
10.21437/Interspeech.2018-1896
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recent advent of deep learning techniques in speech technology and in particular in automatic speech recognition has yielded substantial performance improvements. This suggests that deep neural networks (DNNs) are able to capture structure in speech data that older methods for acoustic modeling, such as Gaussian Mixture Models and shallow neural networks fail to uncover. In image recognition it is possible to link representations on the first couple of layers in DNNs to structural properties of images, and to representations on early layers in the visual cortex. This raises the question whether it is possible to accomplish a similar feat with representations on DNN layers when processing speech input. In this paper we present three different experiments in which we attempt to untangle how DNNs encode speech signals, and to relate these representations to phonetic knowledge, with the aim to advance conventional phonetic concepts and to choose the topology of a DNNs more efficiently. Two experiments investigate representations formed by auto-encoders. A third experiment investigates representations on convolutional layers that treat speech spectrograms as if they were images. The results lay the basis for future experiments with recursive networks.
引用
收藏
页码:1457 / 1461
页数:5
相关论文
共 50 条
  • [21] What Can we Learn (and What Can't we Learn) from Observational Studies of Adolescent Varicocele Treatment?
    Kurtz, Michael P.
    JOURNAL OF UROLOGY, 2017, 198 (01): : 6 - 8
  • [22] Modeling memory: What do we learn from attractor neural networks?
    Brunel, N
    Nadal, JP
    COMPTES RENDUS DE L ACADEMIE DES SCIENCES SERIE III-SCIENCES DE LA VIE-LIFE SCIENCES, 1998, 321 (2-3): : 249 - 252
  • [23] Modeling memory: what do we learn from attractor neural networks?
    Brunel, N.
    Nadal, J.-P.
    Comptes Rendus De L'Academie Des Sciences. Serie III, Sciences De La Vie, 321 (2-3):
  • [24] Neural Encoding for Human Visual Cortex With Deep Neural Networks Learning "What" and "Where"
    Wang, Haibao
    Huang, Lijie
    Du, Changde
    Li, Dan
    Wang, Bo
    He, Huiguang
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2021, 13 (04) : 827 - 840
  • [25] TAYLOR ON HILLSBOROUGH - WHAT CAN WE LEARN
    WALSH, M
    DISASTERS, 1989, 13 (03) : 274 - 277
  • [26] WHAT CAN WE LEARN FROM ART
    DIFFEY, TJ
    AUSTRALASIAN JOURNAL OF PHILOSOPHY, 1995, 73 (02) : 204 - 211
  • [27] Current litigation - what can we learn?
    Harvey, Richard
    EUROPEAN JOURNAL OF PUBLIC HEALTH, 2021, 31
  • [28] WHAT CAN WE LEARN FROM THE PAST
    DEBREE, S
    CHEMISTRY & INDUSTRY, 1995, (23) : 965 - 967
  • [29] Tax federalism: what can we learn?
    Vilalta, Maite
    ACTIVITAT PARLAMENTARIA, 2010, (21): : 23 - 31
  • [30] WHAT CAN WE LEARN FROM THE OCTOPUS?
    Tramacere, F.
    Beccai, L.
    Mazzolai, B.
    BIOLOGICAL AND BIOMIMETIC ADHESIVES: CHALLENGES AND OPPORTUNITIES, 2013, : 89 - 102