Tractable Reinforcement Learning for Signal Temporal Logic Tasks With Counterfactual Experience Replay. Wang, S., Yin, X., Li, S., & Yin, X. IEEE Control Systems Letters, 8:616–621, 2024.
Tractable Reinforcement Learning for Signal Temporal Logic Tasks With Counterfactual Experience Replay [link]Paper  doi  abstract   bibtex   
We investigate the control synthesis problem for Markov decision processes (MDPs) with unknown transition probabilities under signal temporal logic (STL) specifications. Our primary objective is to learn a control policy that maximizes the probability of satisfying the STL task. However, existing approaches to STL control synthesis using reinforcement learning encounter a significant exploration challenge, particularly when expanding the state space to incorporate STL tasks. In this work, we propose a novel reinforcement learning algorithm tailored for STL tasks, addressing the exploration difficulty by effectively leveraging counterfactual experiences to expedite the training process. Through experiments we show that these generated experiences enable us to fully employ the knowledge embedded within the task, resulting in a substantial reduction in the number of trial-and-error explorations required before achieving convergence.
@article{wang_tractable_2024,
	title = {Tractable {Reinforcement} {Learning} for {Signal} {Temporal} {Logic} {Tasks} {With} {Counterfactual} {Experience} {Replay}},
	volume = {8},
	copyright = {https://ieeexplore.ieee.org/Xplorehelp/downloads/license-information/IEEE.html},
	issn = {2475-1456},
	url = {https://ieeexplore.ieee.org/document/10522501/},
	doi = {10.1109/LCSYS.2024.3397844},
	abstract = {We investigate the control synthesis problem for Markov decision processes (MDPs) with unknown transition probabilities under signal temporal logic (STL) specifications. Our primary objective is to learn a control policy that maximizes the probability of satisfying the STL task. However, existing approaches to STL control synthesis using reinforcement learning encounter a significant exploration challenge, particularly when expanding the state space to incorporate STL tasks. In this work, we propose a novel reinforcement learning algorithm tailored for STL tasks, addressing the exploration difficulty by effectively leveraging counterfactual experiences to expedite the training process. Through experiments we show that these generated experiences enable us to fully employ the knowledge embedded within the task, resulting in a substantial reduction in the number of trial-and-error explorations required before achieving convergence.},
	language = {en},
	urldate = {2024-12-18},
	journal = {IEEE Control Systems Letters},
	author = {Wang, Siqi and Yin, Xunyuan and Li, Shaoyuan and Yin, Xiang},
	year = {2024},
	pages = {616--621},
}

Downloads: 0