The Performance Impact of Combining Agent Factorization with Different Learning Algorithms for Multiagent Coordination
Andreas Kallinteris, Stavros Orfanoudakis, Georgios Chalkiadakis
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/stavrosgreece/MultiAgentLearningOfficialpytorch★ 6
Abstract
Factorizing a multiagent system refers to partitioning the state- action space to individual agents and defining the interactions be- tween those agents. This so-called agent factorization is of much im- portance in real-world industrial settings, and is a process that can have significant performance implications. In this work, we explore if the performance impact of agent factorization is different when using different learning algorithms in multiagent coordination set- tings. We evaluated six different agent factorization instances—or agent definitions—in the warehouse traffic management domain, comparing the performance of (mainly) two learning algorithms suitable for learning coordinated multiagent policies: the Evolu- tionary Strategies (ES), and a genetic algorithm (CCEA) previously used in this setting. Our results demonstrate that different learning algorithms are affected in different ways by alternative agent defi- nitions. Given this, we can deduce that many important multiagent coordination problems can potentially be solved by an appropriate agent factorization in conjunction with an appropriate choice of a learning algorithm. Moreover, our work shows that ES is an ef- fective learning algorithm for the warehouse traffic management domain; while, interestingly, celebrated policy gradient methods do not fare well in this complex real-world problem setting.