Backpropagation and stochastic gradient descent method. Amari, S. Neurocomputing, 5(4):185-196, 1993.
doi  abstract   bibtex   
The backpropagation learning method has opened a way to wide applications of neural network research. It is a type of the stochastic descent method known in the sixties. The present paper reviews the wide applicability of the stochastic gradient descent method to various types of models and loss functions. In particular, we apply it to the pattern recognition problem, obtaining a new learning algorithm based on the information criterion. Dynamical properties of learning curves are then studied based on an old paper by the author where the stochastic descent method was proposed for general multilayer networks. The paper is concluded with a short section offering some historical remarks.
@article{SDG,
title = {Backpropagation and stochastic gradient descent method},
journal = {Neurocomputing},
volume = {5},
number = {4},
pages = {185-196},
year = {1993},
issn = {0925-2312},
doi = {https://doi.org/10.1016/0925-2312(93)90006-O},
author = {Shun-ichi Amari},
keywords = {Stochastic descent, generalized delta rule, dynamics of learning, pattern classification, multilayer perceptron},
abstract = {The backpropagation learning method has opened a way to wide applications of neural network research. It is a type of the stochastic descent method known in the sixties. The present paper reviews the wide applicability of the stochastic gradient descent method to various types of models and loss functions. In particular, we apply it to the pattern recognition problem, obtaining a new learning algorithm based on the information criterion. Dynamical properties of learning curves are then studied based on an old paper by the author where the stochastic descent method was proposed for general multilayer networks. The paper is concluded with a short section offering some historical remarks.}
}

Downloads: 0