Short-term Hebbian learning can implement transformer-like attention

被引:2
|
作者
Ellwood, Ian T. [1 ]
机构
[1] Cornell Univ, Dept Neurobiol & Behav, Ithaca, NY 14850 USA
关键词
APICAL DENDRITES; PYRAMIDAL CELLS; SPIKES; MODEL;
D O I
10.1371/journal.pcbi.1011843
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
Transformers have revolutionized machine learning models of language and vision, but their connection with neuroscience remains tenuous. Built from attention layers, they require a mass comparison of queries and keys that is difficult to perform using traditional neural circuits. Here, we show that neurons can implement attention-like computations using short-term, Hebbian synaptic potentiation. We call our mechanism the match-and-control principle and it proposes that when activity in an axon is synchronous, or matched, with the somatic activity of a neuron that it synapses onto, the synapse can be briefly strongly potentiated, allowing the axon to take over, or control, the activity of the downstream neuron for a short time. In our scheme, the keys and queries are represented as spike trains and comparisons between the two are performed in individual spines allowing for hundreds of key comparisons per query and roughly as many keys and queries as there are neurons in the network. Many of the most impressive recent advances in machine learning, from generating images from text to human-like chatbots, are based on a neural network architecture known as the transformer. Transformers are built from so-called attention layers which perform large numbers of comparisons between the vector outputs of the previous layers, allowing information to flow through the network in a more dynamic way than previous designs. This large number of comparisons is computationally expensive and has no known analogue in the brain. Here, we show that a variation on a learning mechanism familiar in neuroscience, Hebbian learning, can implement a transformer-like attention computation if the synaptic weight changes are large and rapidly induced. We call our method the match-and-control principle and it proposes that when presynaptic and postsynaptic spike trains match up, small groups of synapses can be transiently potentiated allowing a few presynaptic axons to control the activity of a neuron. To demonstrate the principle, we build a model of a pyramidal neuron and use it to illustrate the power and limitations of the idea.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Transformer-like model with linear attention for speech emotion recognition
    Du, Jing
    Tang, Manting
    Zhao, Li
    Journal of Southeast University (English Edition), 2021, 37 (02): : 164 - 170
  • [2] LSDASCFormer: A Transformer-Like Network With Long-Short-Distance Attention and SCConv for Heart Rate Measurement
    Liu, Xinhua
    Zhou, Yuheng
    Dong, Zehua
    Ma, Xiaolin
    Kuang, Hailan
    IEEE SENSORS JOURNAL, 2024, 24 (17) : 28007 - 28017
  • [3] Positional Attention Guided Transformer-Like Architecture for Visual Question Answering
    Mao, Aihua
    Yang, Zhi
    Lin, Ken
    Xuan, Jun
    Liu, Yong-Jin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 6997 - 7009
  • [4] An Oscillatory Hebbian Network Model of Short-Term Memory
    Winder, Ransom K.
    Reggia, James A.
    Weems, Scott A.
    Bunting, Michael F.
    NEURAL COMPUTATION, 2009, 21 (03) : 741 - 761
  • [5] Short-term Stress Can Affect Learning and Memory
    严敏
    当代外语研究, 2008, (04) : 8 - 9
  • [6] Long Short-Term Attention
    Zhong, Guoqiang
    Lin, Xin
    Chen, Kang
    Li, Qingyang
    Huang, Kaizhu
    ADVANCES IN BRAIN INSPIRED COGNITIVE SYSTEMS, 2020, 11691 : 45 - 54
  • [7] Describe Molecules by a Heterogeneous Graph Neural Network with Transformer-like Attention for Supervised Property Predictions
    Deng, Daiguo
    Lei, Zengrong
    Hong, Xiaobin
    Zhang, Ruochi
    Zhou, Fengfeng
    ACS OMEGA, 2022, 7 (04): : 3713 - 3721
  • [8] Spatial attention can bias search in visual short-term memory
    Nobre, Anna C.
    Griffin, Ivan C.
    Rao, Anling
    FRONTIERS IN HUMAN NEUROSCIENCE, 2008, 1
  • [9] A Transformer Based Method with Wide Attention Range for Enhanced Short-term Load Forecasting
    Jiang, Bozhen
    Liu, Yi
    Geng, Hua
    Zeng, Huarong
    Ding, Jiangqiao
    2022 4TH INTERNATIONAL CONFERENCE ON SMART POWER & INTERNET ENERGY SYSTEMS, SPIES, 2022, : 1684 - 1690
  • [10] Can Eruptions Be Predicted? Short-Term Prediction of Volcanic Eruptions via Attention-Based Long Short-Term Memory
    Le, Hiep, V
    Murata, Tsuyoshi
    Iguchi, Masato
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13320 - 13325