Stochastic Gradient Descent Tricks. Bottou, L. In Montavon, G., Orr, G. B., & Müller, K., editors, Neural Networks: Tricks of the Trade: Second Edition, of Lecture Notes in Computer Science, pages 421–436. Springer, Berlin, Heidelberg, 2012.
Stochastic Gradient Descent Tricks [link]Paper  doi  abstract   bibtex   
Chapter 1 strongly advocates the stochastic back-propagation method to train neural networks. This is in fact an instance of a more general technique called stochastic gradient descent (SGD). This chapter provides background material, explains why SGD is a good learning algorithm when the training set is large, and provides useful recommendations.
@incollection{bottou_stochastic_2012,
	address = {Berlin, Heidelberg},
	series = {Lecture {Notes} in {Computer} {Science}},
	title = {Stochastic {Gradient} {Descent} {Tricks}},
	isbn = {978-3-642-35289-8},
	url = {https://doi.org/10.1007/978-3-642-35289-8_25},
	abstract = {Chapter 1 strongly advocates the stochastic back-propagation method to train neural networks. This is in fact an instance of a more general technique called stochastic gradient descent (SGD). This chapter provides background material, explains why SGD is a good learning algorithm when the training set is large, and provides useful recommendations.},
	language = {en},
	urldate = {2022-03-19},
	booktitle = {Neural {Networks}: {Tricks} of the {Trade}: {Second} {Edition}},
	publisher = {Springer},
	author = {Bottou, Léon},
	editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
	year = {2012},
	doi = {10.1007/978-3-642-35289-8_25},
	keywords = {Conditional Random Field, Empirical Risk, Learning Rate, Stochastic Gradient, Support Vector Machine},
	pages = {421--436},
}

Downloads: 0