Hebbian dreaming for small datasets

被引:3
|
作者
Agliari, Elena [1 ]
Alemanno, Francesco [4 ]
Aquaro, Miriam [1 ]
Barra, Adriano [4 ]
Durante, Fabrizio [3 ]
Kanter, Ido [2 ]
机构
[1] Sapienza Univ Roma, Dept Math, Rome, Italy
[2] Bar Ilan Univ, Dept Phys, Ramat Gan, Israel
[3] Univ Salento, Dept Econ Sci, Lecce, Italy
[4] Univ Salento, Dept Math & Phys, Lecce, Italy
关键词
Hebbian learning; Sleeping phenomena; Statistical mechanics; Hopfield model; INFORMATION-STORAGE; MEMORY; RETRIEVAL; PATTERNS; SLEEP;
D O I
10.1016/j.neunet.2024.106174
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The dreaming Hopfield model constitutes a generalization of the Hebbian paradigm for neural networks, that is able to perform on -line learning when "awake"and also to account for off -line "sleeping"mechanisms. The latter have been shown to enhance storing in such a way that, in the long sleep-time limit, this model can reach the maximal storage capacity achievable by networks equipped with symmetric pairwise interactions. In this paper, we inspect the minimal amount of information that must be supplied to such a network to guarantee a successful generalization, and we test it both on random synthetic and on standard structured datasets (������.������., MNIST, Fashion-MNIST and Olivetti). By comparing these minimal thresholds of information with those required by the standard (������.������., always "awake") Hopfield model, we prove that the present network can save up to similar to 90% of the dataset size, yet preserving the same performance of the standard counterpart. This suggests that sleep may play a pivotal role in explaining the gap between the large volumes of data required to train artificial neural networks and the relatively small volumes needed by their biological counterparts. Further, we prove that the model Cost function (typically used in statistical mechanics) admits a representation in terms of a standard Loss function (typically used in machine learning) and this allows us to analyze its emergent computational skills both theoretically and computationally: a quantitative picture of its capabilities as a function of its control parameters is achieved and consistency between the two approaches is highlighted. The resulting network is an associative memory for pattern recognition tasks that learns from examples on -line, generalizes correctly (in suitable regions of its control parameters) and optimizes its storage capacity by off -line sleeping: such a reduction of the training cost can be inspiring toward sustainable AI and in situations where data are relatively sparse.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Optimization theory of Hebbian/anti-Hebbian networks for PCA and whitening
    Pehlevan, Cengiz
    Chklovskii, Dmitri B.
    2015 53RD ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2015, : 1458 - 1465
  • [42] PLEA TO STOP DREAMING ABOUT DREAMING
    GALLAGHER, N
    PHILOSOPHY AND PHENOMENOLOGICAL RESEARCH, 1976, 36 (03) : 423 - 424
  • [43] Dreaming as a design tool? ('Dreaming the Landscape')
    Brigham, Don
    LANDSCAPE ARCHITECTURE, 2008, 98 (07): : 19 - 19
  • [44] Smart Cybercrime Classification for Digital Forensics with Small Datasets
    Tuhin, Isfaque Al Kaderi
    Loh, Peter
    Wang, Zhengkui
    CYBER SECURITY, CRYPTOLOGY, AND MACHINE LEARNING, 2022, 13301 : 270 - 280
  • [45] APLYSIA - HEBBIAN OR NOT - REPLY
    SAHLEY, C
    TRENDS IN NEUROSCIENCES, 1986, 9 (09) : 411 - 411
  • [46] Transfer Learning on Small Datasets for Improved Fall Detection
    Maray, Nader
    Ngu, Anne Hee
    Ni, Jianyuan
    Debnath, Minakshi
    Wang, Lu
    SENSORS, 2023, 23 (03)
  • [47] TextCaps : Handwritten Character Recognition with Very Small Datasets
    Jayasundara, Vinoj
    Jayasekara, Sandaru
    Jayasekara, Hirunima
    Rajasegaran, Jathushan
    Seneviratne, Suranga
    Rodrigo, Ranga
    2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, : 254 - 262
  • [48] Improved Vit via knowledge distallation on small datasets
    Wang, Jun
    Liu, Weifeng
    Zhang, Weishan
    Liu, Baodi
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 184 - 189
  • [49] Statistical Prediction of Extreme Events from Small Datasets
    Racca, Alberto
    Magri, Luca
    COMPUTATIONAL SCIENCE - ICCS 2022, PT III, 2022, 13352 : 707 - 713
  • [50] IDSM ChemWebRDF: SPARQLing small-molecule datasets
    Jakub Galgonek
    Jiří Vondrášek
    Journal of Cheminformatics, 13