SOTAVerified

Differentially Private Exploration in Reinforcement Learning with Linear Representation

2021-12-02Unverified0· sign in to hype

Paul Luyo, Evrard Garcelon, Alessandro Lazaric, Matteo Pirotta

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper studies privacy-preserving exploration in Markov Decision Processes (MDPs) with linear representation. We first consider the setting of linear-mixture MDPs (Ayoub et al., 2020) (a.k.a.\ model-based setting) and provide an unified framework for analyzing joint and local differential private (DP) exploration. Through this framework, we prove a O(K^3/4/) regret bound for (,)-local DP exploration and a O(K/) regret bound for (,)-joint DP. We further study privacy-preserving exploration in linear MDPs (Jin et al., 2020) (a.k.a.\ model-free setting) where we provide a O(K^35/^25) regret bound for (,)-joint DP, with a novel algorithm based on low-switching. Finally, we provide insights into the issues of designing local DP algorithms in this model-free setting.

Tasks

Reproductions