Natural Gradient Descent of Complex-Valued Neural Networks Invariant under Rotations

被引:0
|
作者
Mukuno, Jun-ichi [1 ]
Matsui, Hajime [2 ]
机构
[1] Kogakuin Univ, Acad Support Ctr, Hachioji, Tokyo 1920015, Japan
[2] Toyota Technol Inst, Nagoya, Aichi 4688511, Japan
关键词
complex number; Fisher information matrix; projected natural gradient; data augmentation; online character recognition;
D O I
10.1587/transfun.E102.A.1988
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The natural gradient descent is an optimization method for real-valued neural networks that was proposed from the viewpoint of information geometry. Here, we present an extension of the natural gradient descent to complex-valued neural networks. Our idea is to use the Hermitian extension of the Fisher information matrix. Moreover, we generalize the projected natural gradient (PRONG), which is a fast natural gradient descent algorithm, to complex-valued neural networks. We also consider the advantage of complex-valued neural networks over real-valued neural networks. A useful property of complex numbers in the complex plane is that the rotation is simply expressed by the multiplication. By focusing on this property, we construct the output function of complex-valued neural networks, which is invariant even if the input is changed to its rotated value. Then, our complex-valued neural network can learn rotated data without data augmentation. Finally, through simulation of online character recognition, we demonstrate the effectiveness of the proposed approach.
引用
收藏
页码:1988 / 1996
页数:9
相关论文
共 50 条
  • [1] Natural Gradient Descent for Training Stochastic Complex-Valued Neural Networks
    Nitta, Tohru
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2014, 5 (07) : 193 - 198
  • [2] Enhanced Gradient Descent Algorithms for Complex-Valued Neural Networks
    Popa, Calin-Adrian
    16TH INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING (SYNASC 2014), 2014, : 272 - 279
  • [3] Adaptive orthogonal gradient descent algorithm for fully complex-valued neural networks
    Zhao, Weijing
    Huang, He
    NEUROCOMPUTING, 2023, 546
  • [4] Improving the capacity of complex-valued neural networks with a modified gradient descent learning rule
    Lee, DL
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2001, 12 (02): : 439 - 443
  • [5] A data-reusing gradient descent algorithm for complex-valued recurrent neural networks
    Goh, SL
    Mandic, DP
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2003, 2774 : 340 - 350
  • [6] Conjugate Gradient Algorithms for Complex-Valued Neural Networks
    Popa, Calin-Adrian
    NEURAL INFORMATION PROCESSING, PT II, 2015, 9490 : 412 - 422
  • [7] Adaptive stepsize estimation based accelerated gradient descent algorithm for fully complex-valued neural networks
    Zhao, Weijing
    Huang, He
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 236
  • [8] Improving Gradient Regularization using Complex-Valued Neural Networks
    Yeats, Eric
    Chen, Yiran
    Li, Hai
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [9] Scaled Conjugate Gradient Learning for Complex-Valued Neural Networks
    Popa, Calin-Adrian
    MENDEL 2015: RECENT ADVANCES IN SOFT COMPUTING, 2015, 378 : 221 - 233
  • [10] Complex-valued neural networks
    Department of Electrical Engineering and Information Systems, University of Tokyo, 7-3-1, Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
    IEEJ Trans. Electron. Inf. Syst., 1 (2-8):