More Like Real World Game Challenge for Partially Observable Multi-Agent Cooperation

by   Meng Yao, et al.

Some standardized environments have been designed for partially observable multi-agent cooperation, but we find most current environments are synchronous, whereas real-world agents often have their own action spaces leading to asynchrony. Furthermore, fixed agents number limits the scalability of action space, whereas in reality agents number can change resulting in a flexible action space. In addition, current environments are balanced, which is not always the case in the real world where there may be an ability gap between different parties leading to asymmetry. Finally, current environments tend to have less stochasticity with simple state transitions, whereas real-world environments can be highly stochastic and result in extremely risky. To address this gap, we propose WarGame Challenge (WGC) inspired by the Wargame. WGC is a lightweight, flexible, and easy-to-use environment with a clear framework that can be easily configured by users. Along with the benchmark, we provide MARL baseline algorithms such as QMIX and a toolkit to help algorithms complete performance tests on WGC. Finally, we present baseline experiment results, which demonstrate the challenges of WGC. We think WGC enrichs the partially observable multi-agent cooperation domain and introduces more challenges that better reflect the real-world characteristics. Code is release in


page 2

page 3

page 7


The Partially Observable Asynchronous Multi-Agent Cooperation Challenge

Multi-agent reinforcement learning (MARL) has received increasing attent...

Soft Hierarchical Graph Recurrent Networks for Many-Agent Partially Observable Environments

The recent progress in multi-agent deep reinforcement learning(MADRL) ma...

POGEMA: Partially Observable Grid Environment for Multiple Agents

We introduce POGEMA ( a sandbox...

Q-Mixing Network for Multi-Agent Pathfinding in Partially Observable Grid Environments

In this paper, we consider the problem of multi-agent navigation in part...

Context Discovery for Model Learning in Partially Observable Environments

The ability to learn a model is essential for the success of autonomous ...

Multi-Agent Intention Sharing via Leader-Follower Forest

Intention sharing is crucial for efficient cooperation under partially o...

R-MADDPG for Partially Observable Environments and Limited Communication

There are several real-world tasks that would ben-efit from applying mul...

Please sign up or login with your details

Forgot password? Click here to reset