SOTAVerified

How are policy gradient methods affected by the limits of control?

2022-06-14Unverified0· sign in to hype

Ingvar Ziemann, Anastasios Tsiamis, Henrik Sandberg, Nikolai Matni

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We study stochastic policy gradient methods from the perspective of control-theoretic limitations. Our main result is that ill-conditioned linear systems in the sense of Doyle inevitably lead to noisy gradient estimates. We also give an example of a class of stable systems in which policy gradient methods suffer from the curse of dimensionality. Our results apply to both state feedback and partially observed systems.

Tasks

Reproductions