Learning to learn using gradient descent

被引:0
|
作者
Hochreiter, S [1 ]
Younger, AS
Conwell, PR
机构
[1] Univ Colorado, Dept Comp Sci, Boulder, CO 80309 USA
[2] Westminster Coll, Dept Phys, Salt Lake City, UT USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper introduces the application of gradient descent methods to meta-learning. The concept of "meta-learning", i.e. of a system that improves or discovers a learning algorithm, has been of interest in machine learning for decades because of its appealing applications. Previous meta-learning approaches have been based on evolutionary methods and, therefore, have been restricted to small models with few free parameters. We make meta-learning in large systems feasible by using recurrent neural networks with their attendant learning routines as meta-learning systems. Our system derived complex well performing learning algorithms from scratch. In this paper we also show that our approach performs non-stationary time series prediction.
引用
收藏
页码:87 / 94
页数:8
相关论文
共 50 条
  • [41] Robust supervised learning with coordinate gradient descent
    Ibrahim Merad
    Stéphane Gaïffas
    [J]. Statistics and Computing, 2023, 33
  • [42] Natural gradient descent for on-line learning
    [J]. Phys Rev Lett, 24 (5461):
  • [43] Learning with Gradient Descent and Weakly Convex Losses
    Richards, Dominic
    Rabbat, Mike
    [J]. 24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [44] Learning rates of gradient descent algorithm for classification
    Dong, Xue-Mei
    Chen, Di-Rong
    [J]. JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2009, 224 (01) : 182 - 192
  • [45] Robust supervised learning with coordinate gradient descent
    Merad, Ibrahim
    Gaiffas, Stephane
    [J]. STATISTICS AND COMPUTING, 2023, 33 (05)
  • [46] Deep learning for sea cucumber detection using stochastic gradient descent algorithm
    Zhang, Huaqiang
    Yu, Fusheng
    Sun, Jincheng
    Shen, Xiaoqin
    Li, Kun
    [J]. EUROPEAN JOURNAL OF REMOTE SENSING, 2020, 53 (53-62) : 53 - 62
  • [47] Hopfield neural network learning using direct gradient descent of energy function
    Tang, Z
    Tashima, K
    Hebishima, H
    Ishizuka, O
    Tanno, K
    [J]. IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 1996, E79A (02) : 258 - 261
  • [48] Hopfield neural network learning using direct gradient descent of energy function
    Miyazaki Univ, Miyazaki-shi, Japan
    [J]. IEICE Trans Fund Electron Commun Comput Sci, 2 (258-261):
  • [49] From Gradient Flow on Population Loss to Learning with Stochastic Gradient Descent
    Sekhari, Ayush
    Kale, Satyen
    Lee, Jason D.
    De Sa, Chris
    Sridharan, Karthik
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [50] How to Learn when Data Reacts to Your Model: Performative Gradient Descent
    Izzo, Zachary
    Ying, Lexing
    Zou, James
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139