SOTAVerified

Asynchronous, Option-Based Multi-Agent Policy Gradient: A Conditional Reasoning Approach

2022-03-29Unverified0· sign in to hype

Xubo Lyu, Amin Banitalebi-Dehkordi, Mo Chen, Yong Zhang

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Cooperative multi-agent problems often require coordination between agents, which can be achieved through a centralized policy that considers the global state. Multi-agent policy gradient (MAPG) methods are commonly used to learn such policies, but they are often limited to problems with low-level action spaces. In complex problems with large state and action spaces, it is advantageous to extend MAPG methods to use higher-level actions, also known as options, to improve the policy search efficiency. However, multi-robot option executions are often asynchronous, that is, agents may select and complete their options at different time steps. This makes it difficult for MAPG methods to derive a centralized policy and evaluate its gradient, as centralized policy always select new options at the same time. In this work, we propose a novel, conditional reasoning approach to address this problem and demonstrate its effectiveness on representative option-based multi-agent cooperative tasks through empirical validation. Find code and videos at: https://sites.google.com/view/mahrlsupp/

Tasks

Reproductions