Jaynes machine: The universal microstructure of deep neural networks

被引:1
|
作者
Venkatasubramanian, Venkat [1 ]
Sanjeevrajan, N. [1 ]
Khandekar, Manasi [2 ]
Sivaram, Abhishek [3 ]
Szczepanski, Collin [1 ]
机构
[1] Columbia Univ, Dept Chem Engn, Complex Resilient Intelligent Syst Lab, New York, NY 10027 USA
[2] Columbia Univ, Dept Comp Sci & Engn, New York, NY 10027 USA
[3] Tech Univ Denmark, Dept Chem & Biochem Engn, DK-2800 Lyngby, Denmark
关键词
LLMs; Boltzmann machine; Hopfield networks; Game theory; Arbitrage equilibrium; Deep learning; DESIGN; SYSTEMS;
D O I
10.1016/j.compchemeng.2024.108908
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Despite the recent stunning progress in large-scale deep neural network applications, our understanding of their microstructure, 'energy' functions, and optimal design remains incomplete. Here, we present anew game- theoretic framework, called statistical teleodynamics, that reveals important insights into these key properties. The optimally robust design of such networks inherently involves computational benefit-cost trade-offs that physics-inspired models do not adequately capture. These trade-offs occur as neurons and connections compete to increase their effective utilities under resource constraints during training. Ina fully trained network, this results in a state of arbitrage equilibrium, where all neurons in a given layer have the same effective utility, and all connections to a given layer have the same effective utility. The equilibrium is characterized by the emergence of two lognormal distributions of connection weights and neuronal output as the universal microstructure of large deep neural networks. We call such a network the Jaynes Machine. Our theoretical predictions are shown to be supported by empirical data from seven large-scale deep neural networks. We also show that the Hopfield network and the Boltzmann Machine are the same special case of the Jaynes Machine.
引用
收藏
页数:10
相关论文
共 50 条
  • [11] Deep Neural Networks and Explainable Machine Learning
    Maratea, Antonio
    Ferone, Alessio
    FUZZY LOGIC AND APPLICATIONS, WILF 2018, 2019, 11291 : 253 - 256
  • [12] Deep neural networks watermark via universal deep hiding and metric learning
    Zhicheng Ye
    Xinpeng Zhang
    Guorui Feng
    Neural Computing and Applications, 2024, 36 : 7421 - 7438
  • [13] Deep neural networks watermark via universal deep hiding and metric learning
    Ye, Zhicheng
    Zhang, Xinpeng
    Feng, Guorui
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (13): : 7421 - 7438
  • [14] Universal backdoor attack on deep neural networks for malware detection
    Zhang, Yunchun
    Feng, Fan
    Liao, Zikun
    Li, Zixuan
    Yao, Shaowen
    APPLIED SOFT COMPUTING, 2023, 143
  • [15] Deep Neural Networks, Generic Universal Interpolation, and Controlled ODEs
    Cuchiero, Christa
    Larsson, Martin
    Teichmann, Josef
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (03): : 901 - 919
  • [16] GRAPH EXPANSIONS OF DEEP NEURAL NETWORKS AND THEIR UNIVERSAL SCALING LIMITS
    Cirone, Nicola Muça
    Hamdan, Jad
    Salvi, Cristopher
    arXiv,
  • [17] A Universal VAD Based on Jointly Trained Deep Neural Networks
    Wang, Qing
    Du, Jun
    Bao, Xiao
    Wang, Zi-Rui
    Dai, Li-Rong
    Lee, Chin-Hui
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2282 - 2286
  • [18] A universal strategy for smoothing deceleration in deep graph neural networks
    Cheng, Qi
    Long, Lang
    Xu, Jiayu
    Zhang, Min
    Han, Shuangze
    Zhao, Chengkui
    Feng, Weixing
    NEURAL NETWORKS, 2025, 185
  • [19] Introduction to Machine Learning, Neural Networks, and Deep Learning
    Choi, Rene Y.
    Coyner, Aaron S.
    Kalpathy-Cramer, Jayashree
    Chiang, Michael F.
    Campbell, J. Peter
    TRANSLATIONAL VISION SCIENCE & TECHNOLOGY, 2020, 9 (02):
  • [20] Deep Complex Convolutional Neural Networks for Subwavelength Microstructure Imaging
    Wei, Teng-Fei
    Wang, Xiao-Hua
    Qu, Cheng-Hui
    IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, 2022, 70 (08) : 6329 - 6335