Online and Robust Intermittent Motion Planning in Dynamic and Changing Environments. Xu, Z., Kontoudis, G. P, & Vamvoudakis, K. G IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2023. Pdf Video doi abstract bibtex 14 downloads In this paper, we propose RRT-Q$^{\textrm{X}}_{∞}$, an online and intermittent kinodynamic motion planning framework for dynamic environments with unknown robot dynamics and unknown disturbances. We leverage RRT$^{\textrm{X}}$ for global path planning and rapid replanning to produce waypoints as a sequence of boundary value problems (BVPs). For each BVP, we formulate a finite-horizon, continuous-time zero-sum game, where the control input is the minimizer, and the worst-case disturbance is the maximizer. We propose a robust intermittent Q-learning controller for waypoint navigation with completely unknown system dynamics, external disturbances, and intermittent control updates. We execute a relaxed persistence of excitation technique to guarantee that the Q-learning controller converges to the optimal controller. We provide rigorous Lyapunov-based proofs to guarantee the closed-loop stability of the equilibrium point. The effectiveness of the proposed RRT-Q$^{\textrm{X}}_{∞}$ is illustrated with Monte-Carlo numerical experiments in numerous dynamic and changing environments.
@article{xu2023TNNLS,
title={Online and Robust Intermittent Motion Planning in Dynamic and Changing Environments},
abstract = {In this paper, we propose RRT-Q$^{\textrm{X}}_{\infty}$, an online and intermittent kinodynamic motion planning framework for dynamic environments with unknown robot dynamics and unknown disturbances. We leverage RRT$^{\textrm{X}}$ for global path planning and rapid replanning to produce waypoints as a sequence of boundary value problems (BVPs). For each BVP, we formulate a finite-horizon, continuous-time zero-sum game, where the control input is the minimizer, and the worst-case disturbance is the maximizer. We propose a \textit{robust intermittent Q-learning} controller for waypoint navigation with completely unknown system dynamics, external disturbances, and intermittent control updates. We execute a relaxed persistence of excitation technique to guarantee that the Q-learning controller converges to the optimal controller. We provide rigorous Lyapunov-based proofs to guarantee the closed-loop stability of the equilibrium point. The effectiveness of the proposed RRT-Q$^{\textrm{X}}_{\infty}$ is illustrated with Monte-Carlo numerical experiments in numerous dynamic and changing environments.},
author={Xu, Zirui and Kontoudis, George P and Vamvoudakis, Kyriakos G},
journal={IEEE Transactions on Neural Networks and Learning Systems (TNNLS)},
year={2023},
keywords={motion planning, reinforcement learning, optimal control, event-trigger control, game theory},
url_pdf= {/publications/TNNLS23_Xu_OnlineRobustIntermittentMotionPlanningDynamicChangingEnvironments.pdf},
url_video = {https://youtu.be/iS_PzDmlpfs},
doi = {10.1109/TNNLS.2023.3303811}
}
Downloads: 14
{"_id":"nWFdSpG6YK7Pt7Yth","bibbaseid":"xu-kontoudis-vamvoudakis-onlineandrobustintermittentmotionplanningindynamicandchangingenvironments-2023","author_short":["Xu, Z.","Kontoudis, G. P","Vamvoudakis, K. G"],"bibdata":{"bibtype":"article","type":"article","title":"Online and Robust Intermittent Motion Planning in Dynamic and Changing Environments","abstract":"In this paper, we propose RRT-Q$^{\\textrm{X}}_{∞}$, an online and intermittent kinodynamic motion planning framework for dynamic environments with unknown robot dynamics and unknown disturbances. We leverage RRT$^{\\textrm{X}}$ for global path planning and rapid replanning to produce waypoints as a sequence of boundary value problems (BVPs). For each BVP, we formulate a finite-horizon, continuous-time zero-sum game, where the control input is the minimizer, and the worst-case disturbance is the maximizer. We propose a <i>robust intermittent Q-learning</i> controller for waypoint navigation with completely unknown system dynamics, external disturbances, and intermittent control updates. We execute a relaxed persistence of excitation technique to guarantee that the Q-learning controller converges to the optimal controller. We provide rigorous Lyapunov-based proofs to guarantee the closed-loop stability of the equilibrium point. The effectiveness of the proposed RRT-Q$^{\\textrm{X}}_{∞}$ is illustrated with Monte-Carlo numerical experiments in numerous dynamic and changing environments.","author":[{"propositions":[],"lastnames":["Xu"],"firstnames":["Zirui"],"suffixes":[]},{"propositions":[],"lastnames":["Kontoudis"],"firstnames":["George","P"],"suffixes":[]},{"propositions":[],"lastnames":["Vamvoudakis"],"firstnames":["Kyriakos","G"],"suffixes":[]}],"journal":"IEEE Transactions on Neural Networks and Learning Systems (TNNLS)","year":"2023","keywords":"motion planning, reinforcement learning, optimal control, event-trigger control, game theory","url_pdf":"/publications/TNNLS23_Xu_OnlineRobustIntermittentMotionPlanningDynamicChangingEnvironments.pdf","url_video":"https://youtu.be/iS_PzDmlpfs","doi":"10.1109/TNNLS.2023.3303811","bibtex":"@article{xu2023TNNLS,\n title={Online and Robust Intermittent Motion Planning in Dynamic and Changing Environments},\n abstract = {In this paper, we propose RRT-Q$^{\\textrm{X}}_{\\infty}$, an online and intermittent kinodynamic motion planning framework for dynamic environments with unknown robot dynamics and unknown disturbances. We leverage RRT$^{\\textrm{X}}$ for global path planning and rapid replanning to produce waypoints as a sequence of boundary value problems (BVPs). For each BVP, we formulate a finite-horizon, continuous-time zero-sum game, where the control input is the minimizer, and the worst-case disturbance is the maximizer. We propose a \\textit{robust intermittent Q-learning} controller for waypoint navigation with completely unknown system dynamics, external disturbances, and intermittent control updates. We execute a relaxed persistence of excitation technique to guarantee that the Q-learning controller converges to the optimal controller. We provide rigorous Lyapunov-based proofs to guarantee the closed-loop stability of the equilibrium point. The effectiveness of the proposed RRT-Q$^{\\textrm{X}}_{\\infty}$ is illustrated with Monte-Carlo numerical experiments in numerous dynamic and changing environments.},\n author={Xu, Zirui and Kontoudis, George P and Vamvoudakis, Kyriakos G},\n journal={IEEE Transactions on Neural Networks and Learning Systems (TNNLS)},\n year={2023},\n keywords={motion planning, reinforcement learning, optimal control, event-trigger control, game theory},\n url_pdf= {/publications/TNNLS23_Xu_OnlineRobustIntermittentMotionPlanningDynamicChangingEnvironments.pdf},\n url_video = {https://youtu.be/iS_PzDmlpfs},\n doi = {10.1109/TNNLS.2023.3303811}\n}\n\n","author_short":["Xu, Z.","Kontoudis, G. P","Vamvoudakis, K. G"],"key":"xu2023TNNLS","id":"xu2023TNNLS","bibbaseid":"xu-kontoudis-vamvoudakis-onlineandrobustintermittentmotionplanningindynamicandchangingenvironments-2023","role":"author","urls":{" pdf":"http://www.georgekontoudis.com/publications/TNNLS23_Xu_OnlineRobustIntermittentMotionPlanningDynamicChangingEnvironments.pdf"," video":"https://youtu.be/iS_PzDmlpfs"},"keyword":["motion planning","reinforcement learning","optimal control","event-trigger control","game theory"],"metadata":{"authorlinks":{}},"downloads":14,"html":""},"bibtype":"article","biburl":"http://www.georgekontoudis.com/publications/Kontoudis_mypubs.bib","dataSources":["ENfCwQrwsiuEJtsqK","ya2CyA73rpZseyrZ8","2252seNhipfTmjEBQ"],"keywords":["motion planning","reinforcement learning","optimal control","event-trigger control","game theory"],"search_terms":["online","robust","intermittent","motion","planning","dynamic","changing","environments","xu","kontoudis","vamvoudakis"],"title":"Online and Robust Intermittent Motion Planning in Dynamic and Changing Environments","year":2023,"downloads":14}