Skip to yearly menu bar Skip to main content

In-Person Poster presentation / poster accept

Provable Sim-to-real Transfer in Continuous Domain with Partial Observations

Jiachen Hu · Han Zhong · Chi Jin · Liwei Wang

MH1-2-3-4 #101

Keywords: [ Reinforcement Learning ] [ partial observations ] [ sim-to-real ] [ RL Theory ]

Sim-to-real transfer, which trains RL agents in the simulated environments and then deploys them in the real world, has been widely used to overcome the limitations of gathering samples in the real world. Despite the empirical success of the sim-to-real transfer, its theoretical foundation is much less understood. In this paper, we study the sim-to-real transfer in continuous domain with partial observations, where the simulated environments and real-world environments are modeled by linear quadratic Gaussian (LQG) systems. We show that a popular robust adversarial training algorithm is capable of learning a policy from the simulated environment that is competitive to the optimal policy in the real-world environment. To achieve our results, we design a new algorithm for infinite-horizon average-cost LQGs and establish a regret bound that depends on the intrinsic complexity of the model class. Our algorithm crucially relies on a novel history clipping scheme, which might be of independent interest.

Chat is not available.