SOTAVerified

Exploiting Symmetry in Dynamics for Model-Based Reinforcement Learning with Asymmetric Rewards

2024-03-27Code Available0· sign in to hype

Yasin Sonmez, Neelay Junnarkar, Murat Arcak

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Recent work in reinforcement learning has leveraged symmetries in the model to improve sample efficiency in training a policy. A commonly used simplifying assumption is that the dynamics and reward both exhibit the same symmetry; however, in many real-world environments, the dynamical model exhibits symmetry independent of the reward model. In this paper, we assume only the dynamics exhibit symmetry, extending the scope of problems in reinforcement learning and learning in control theory to which symmetry techniques can be applied. We use Cartan's moving frame method to introduce a technique for learning dynamics that, by construction, exhibit specified symmetries. Numerical experiments demonstrate that the proposed method learns a more accurate dynamical model

Tasks

Reproductions