Safe Multi-Agent Reinforcement Learning via Shielding

01/27/2021
by   Ingy Elsayed-Aly, et al.
0

Multi-agent reinforcement learning (MARL) has been increasingly used in a wide range of safety-critical applications, which require guaranteed safety (e.g., no unsafe states are ever visited) during the learning process.Unfortunately, current MARL methods do not have safety guarantees. Therefore, we present two shielding approaches for safe MARL. In centralized shielding, we synthesize a single shield to monitor all agents' joint actions and correct any unsafe action if necessary. In factored shielding, we synthesize multiple shields based on a factorization of the joint state space observed by all agents; the set of shields monitors agents concurrently and each shield is only responsible for a subset of agents at each step.Experimental results show that both approaches can guarantee the safety of agents during learning without compromising the quality of learned policies; moreover, factored shielding is more scalable in the number of agents than centralized shielding.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2021

Safe Multi-Agent Reinforcement Learning through Decentralized Multiple Control Barrier Functions

Multi-Agent Reinforcement Learning (MARL) algorithms show amazing perfor...
research
04/13/2023

Model-based Dynamic Shielding for Safe and Efficient Multi-Agent Reinforcement Learning

Multi-Agent Reinforcement Learning (MARL) discovers policies that maximi...
research
02/02/2021

An Abstraction-based Method to Check Multi-Agent Deep Reinforcement-Learning Behaviors

Multi-agent reinforcement learning (RL) often struggles to ensure the sa...
research
04/10/2017

Dynamic Safe Interruptibility for Decentralized Multi-Agent Reinforcement Learning

In reinforcement learning, agents learn by performing actions and observ...
research
04/16/2020

MARLeME: A Multi-Agent Reinforcement Learning Model Extraction Library

Multi-Agent Reinforcement Learning (MARL) encompasses a powerful class o...
research
10/06/2020

Safety Aware Reinforcement Learning (SARL)

As reinforcement learning agents become increasingly integrated into com...
research
12/17/2020

Online Shielding for Stochastic Systems

In this paper, we propose a method to develop trustworthy reinforcement ...

Please sign up or login with your details

Forgot password? Click here to reset