Overview Statistic: PDF-Downloads (blue) and Frontdoor-Views (gray)

Connecting Stochastic Optimal Control and Reinforcement Learning

under review
  • In this article we study the connection of stochastic optimal control and reinforcement learning. Our main motivation is an importance sampling application to rare events sampling which can be reformulated as an optimal control problem. By using a parameterized approach the optimal control problem turns into a stochastic optimization problem which still presents some open questions regarding how to tackle the scalability to high-dimensional problems and how to deal with the intrinsic metastability of the system. With the aim to explore new methods we connect the optimal control problem to reinforcement learning since both share the same underlying framework namely a Markov decision process (MDP). We show how the MDP can be formulated for the optimal control problem. Furthermore, we discuss how the stochastic optimal control problem can be interpreted in a reinforcement learning framework. At the end of the article we present the application of two different reinforcement learning algorithms to the optimal control problem and compare the advantages and disadvantages of the two algorithms.
Metadaten
Author:Jannes Quer, Enric Ribera BorrellORCiD
Document Type:Article
Parent Title (English):Journal of Mathematical Physics
Year of first publication:2022
ArXiv Id:http://arxiv.org/abs/2211.02474
Accept ✔
Diese Webseite verwendet technisch erforderliche Session-Cookies. Durch die weitere Nutzung der Webseite stimmen Sie diesem zu. Unsere Datenschutzerklärung finden Sie hier.