Following Newton direction in Policy Gradient with parameter exploration. Manganini, G., Pirotta, M., Restelli, M., & Bascetta, L. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), pages 1--8, 2015. Institute of Electrical and Electronics Engineers Inc..
doi  abstract   bibtex   
This paper investigates the use of second-order methods to solve Markov Decision Processes (MDPs). Despite the popularity of second-order methods in optimization literature, so far little attention has been paid to the extension of such techniques to face sequential decision problems. Here we provide a model-free Reinforcement Learning method that estimates the Newton direction by sampling directly in the parameter space. In order to compute the Newton direction we provide the formulation of the Hessian of the expected return, a technique for variance reduction in the sample-based estimation and a finite sample analysis in the case of Normal distribution. Beside discussing the theoretical properties, we empirically evaluate the method on an instructional linear-quadratic regulator and on a complex dynamical quadrotor system.
@inproceedings{manganini_following_2015,
	title = {Following {Newton} direction in {Policy} {Gradient} with parameter exploration},
	doi = {10.1109/IJCNN.2015.7280673},
	abstract = {This paper investigates the use of second-order methods to solve Markov Decision Processes (MDPs). Despite the popularity of second-order methods in optimization literature, so far little attention has been paid to the extension of such techniques to face sequential decision problems. Here we provide a model-free Reinforcement Learning method that estimates the Newton direction by sampling directly in the parameter space. In order to compute the Newton direction we provide the formulation of the Hessian of the expected return, a technique for variance reduction in the sample-based estimation and a finite sample analysis in the case of Normal distribution. Beside discussing the theoretical properties, we empirically evaluate the method on an instructional linear-quadratic regulator and on a complex dynamical quadrotor system.},
	booktitle = {Proceedings of the {International} {Joint} {Conference} on {Neural} {Networks} ({IJCNN})},
	publisher = {Institute of Electrical and Electronics Engineers Inc.},
	author = {Manganini, Giorgio and Pirotta, Matteo and Restelli, Marcello and Bascetta, Luca},
	year = {2015},
	keywords = {AUT, INF},
	pages = {1--8}
}

Downloads: 0