共 10 条
- [2] Riemannian metrics for neural networks I: feedforward networks[J] . Ollivier Yann. Information and Inference: A Journal of the IMA . 2015 (2)
- [3] Optimizing Neural Networks with Kronecker-factored Approximate Curvature.[J] . James Martens,Roger B. Grosse. CoRR . 2015
- [4] Adaptive Subgradient Methods for Online Learning and Stochastic Optimization.[J] . John C. Duchi,Elad Hazan,Yoram Singer. Journal of Machine Learning Research . 2011
- [5] On the momentum term in gradient descent learning algorithms [J]. NEURAL NETWORKS, 1999, 12 (01) : 145 - 151
- [6] Natural gradient works efficiently in learning [J]. NEURAL COMPUTATION, 1998, 10 (02) : 251 - 276
- [7] USE OF A QUASI-NEWTON METHOD IN A FEEDFORWARD NEURAL-NETWORK CONSTRUCTION ALGORITHM [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1995, 6 (01): : 273 - 277
- [8] On the limited memory BFGS method for large scale optimization[J] . Dong C. Liu,Jorge Nocedal. Mathematical Programming . 1989 (1-3)
- [9] Large-scale distributed second-order optimization using kronecker-factored approximate curvature for deep convolutional neural networks .2 Osawa K,Tsuji Y,Ueno Y,et al. Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition . 2019
- [10] Learning multiple layers of features from tiny images .2 Krizhevsky,A. University ofToronto . 2009