Skip to yearly menu bar Skip to main content


Poster

Rich-Observation Reinforcement Learning with Continuous Latent Dynamics

Yuda Song · Lili Wu · Dylan Foster · Akshay Krishnamurthy


Abstract:

Sample-efficiency and reliability remain major bottlenecks toward wide adoption of reinforcement learning algorithms in continuous settings with high-dimensional perceptual inputs. Toward addressing these challenges, we introduce a new theoretical framework, RichCLD (“Rich-Observation RL with Continuous Latent Dynamics”), in which the agent performs control based on high-dimensional observations, but the environment is governed by low-dimensional latent states and Lipschitz continuous dynamics. Our main contribution is a new provably statistically and computationally efficient algorithm for this setting. The core of our algorithm is a new representation learning objective; we show that prior representation learning schemes tailored to discrete dynamics do not naturally extend to the continuous setting. Our new objective is amenable to practical implementation, and empirically, we find that it compares favorably to prior schemes.

Live content is unavailable. Log in and register to view live content