Logical Team Q-learning: An approach towards factored policies in cooperative MARL

06/05/2020
by   Lucas Cassano, et al.
0

We address the challenge of learning factored policies in cooperative MARL scenarios. In particular, we consider the situation in which a team of agents collaborates to optimize a common cost. Our goal is to obtain factored policies that determine the individual behavior of each agent so that the resulting joint policy is optimal. In this work we make contributions to both the dynamic programming and reinforcement learning settings. In the dynamic programming case we provide a number of lemmas that prove the existence of such factored policies and we introduce an algorithm (along with proof of convergence) that provably leads to them. Then we introduce tabular and deep versions of Logical Team Q-learning, which is a stochastic version of the algorithm for the RL case. We conclude the paper by providing experiments that illustrate the claims.

READ FULL TEXT
research
05/16/2021

Optimal control of robust team stochastic games

In stochastic dynamic environments, team stochastic games have emerged a...
research
11/02/2017

A topology for Team Policies and Existence of Optimal Team Policies in Stochastic Team Theory

In this paper, we establish the existence of team-optimal policies for s...
research
08/28/2023

Policy Diversity for Cooperative Agents

Standard cooperative multi-agent reinforcement learning (MARL) methods a...
research
07/03/2019

Safe Approximate Dynamic Programming Via Kernelized Lipschitz Estimation

We develop a method for obtaining safe initial policies for reinforcemen...
research
05/17/2019

Non-signaling Approximations of Stochastic Team Problems

In this paper, we consider non-signaling approximation of finite stochas...
research
09/12/2017

Information Design in Crowdfunding under Thresholding Policies

In crowdfunding, an entrepreneur often has to decide how to disclose the...

Please sign up or login with your details

Forgot password? Click here to reset