SOTAVerified

A Large Deviations Perspective on Policy Gradient Algorithms

2023-11-13Unverified0· sign in to hype

Wouter Jongeneel, Daniel Kuhn, Mengmeng Li

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Motivated by policy gradient methods in the context of reinforcement learning, we identify a large deviation rate function for the iterates generated by stochastic gradient descent for possibly non-convex objectives satisfying a Polyak- ojasiewicz condition. Leveraging the contraction principle from large deviations theory, we illustrate the potential of this result by showing how convergence properties of policy gradient with a softmax parametrization and an entropy regularized objective can be naturally extended to a wide spectrum of other policy parametrizations.

Tasks

Reproductions