Copycat perceptron: Smashing barriers through collective learning

被引:0
|
作者
Catania, Giovanni [1 ]
Decelle, Aurelien [1 ,2 ]
Seoane, Beatriz [2 ]
机构
[1] Univ Complutense Madrid, Dept Fis Teor, Madrid 28040, Spain
[2] Univ Paris Saclay, Inria TAU Team, CNRS, LISN, F-91190 Gif Sur yvette, France
关键词
STATISTICAL-MECHANICS; NEURAL-NETWORK; TRANSITION; DYNAMICS; STORAGE;
D O I
10.1103/PhysRevE.109.065313
中图分类号
O35 [流体力学]; O53 [等离子体物理学];
学科分类号
070204 ; 080103 ; 080704 ;
摘要
We characterize the equilibrium properties of a model of y coupled binary perceptrons in the teacher-student scenario, subject to a suitable cost function, with an explicit ferromagnetic coupling proportional to the Hamming distance between the students' weights. In contrast to recent works, we analyze a more general setting in which thermal noise is present that affects each student's generalization performance. In the nonzero temperature regime, we find that the coupling of replicas leads to a bend of the phase diagram towards smaller values of alpha: This suggests that the free entropy landscape gets smoother around the solution with perfect generalization (i.e., the teacher) at a fixed fraction of examples, allowing standard thermal updating algorithms such as Simulated Annealing to easily reach the teacher solution and avoid getting trapped in metastable states as happens in the unreplicated case, even in the computationally easy regime of the inference phase diagram. These results provide additional analytic and numerical evidence for the recently conjectured Bayes-optimal property of Replicated Simulated Annealing for a sufficient number of replicas. From a learning perspective, these results also suggest that multiple students working together (in this case reviewing the same data) are able to learn the same rule both significantly faster and with fewer examples, a property that could be exploited in the context of cooperative and federated learning.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Semidefinite programming by perceptron learning
    Graepel, T
    Herbrich, R
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 16, 2004, 16 : 457 - 464
  • [32] ON THE RATE OF CONVERGENCE OF PERCEPTRON LEARNING
    BHATTACHARYA, U
    PARUI, SK
    PATTERN RECOGNITION LETTERS, 1995, 16 (05) : 491 - 497
  • [33] Accelerating kernel Perceptron learning
    Garcia, Daniel
    Gonzalez, Ana
    Dorronsoro, Jose R.
    ARTIFICIAL NEURAL NETWORKS - ICANN 2007, PT 1, PROCEEDINGS, 2007, 4668 : 159 - +
  • [34] Learning strategy for the binary perceptron
    Reimers, L
    Bouten, M
    VanRompaey, B
    JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1996, 29 (19): : 6247 - 6252
  • [35] New development: Learning communities-an approach to dismantling barriers to collective improvement
    Wilson, Louise
    Hawkins, Melissa
    French, Max
    Lowe, Toby
    Hesselgreaves, Hannah
    PUBLIC MONEY & MANAGEMENT, 2023, 43 (04) : 374 - 377
  • [36] LEARNING WITH NOISE IN A LINEAR PERCEPTRON
    KROGH, A
    JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1992, 25 (05): : 1119 - 1133
  • [37] Perceptron learning in the domain of graphs
    Jain, BJ
    Wysotzki, F
    PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS 2003, VOLS 1-4, 2003, : 1993 - 1998
  • [38] A THERMAL PERCEPTRON LEARNING RULE
    FREAN, M
    NEURAL COMPUTATION, 1992, 4 (06) : 946 - 957
  • [39] LEARNING BY DIFFUSION FOR MULTILAYER PERCEPTRON
    HOPTROFF, RG
    HALL, TJ
    ELECTRONICS LETTERS, 1989, 25 (08) : 531 - 533
  • [40] Unsupervised learning of sigmoid perceptron
    Uykan, Z
    Koivo, HN
    2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS, VOLS I-VI, 2000, : 3486 - 3489