Adaptive Agent Architecture for Real-time Human-Agent Teaming

by   Tianwei Ni, et al.

Teamwork is a set of interrelated reasoning, actions and behaviors of team members that facilitate common objectives. Teamwork theory and experiments have resulted in a set of states and processes for team effectiveness in both human-human and agent-agent teams. However, human-agent teaming is less well studied because it is so new and involves asymmetry in policy and intent not present in human teams. To optimize team performance in human-agent teaming, it is critical that agents infer human intent and adapt their polices for smooth coordination. Most literature in human-agent teaming builds agents referencing a learned human model. Though these agents are guaranteed to perform well with the learned model, they lay heavy assumptions on human policy such as optimality and consistency, which is unlikely in many real-world scenarios. In this paper, we propose a novel adaptive agent architecture in human-model-free setting on a two-player cooperative game, namely Team Space Fortress (TSF). Previous human-human team research have shown complementary policies in TSF game and diversity in human players' skill, which encourages us to relax the assumptions on human policy. Therefore, we discard learning human models from human data, and instead use an adaptation strategy on a pre-trained library of exemplar policies composed of RL algorithms or rule-based methods with minimal assumptions of human behavior. The adaptation strategy relies on a novel similarity metric to infer human policy and then selects the most complementary policy in our library to maximize the team performance. The adaptive agent architecture can be deployed in real-time and generalize to any off-the-shelf static agents. We conducted human-agent experiments to evaluate the proposed adaptive agent framework, and demonstrated the suboptimality, diversity, and adaptability of human policies in human-agent teams.


Policy Diversity for Cooperative Agents

Standard cooperative multi-agent reinforcement learning (MARL) methods a...

The Harmony Index: a Utilitarian Metric for Measuring Effectiveness in Mixed-Skill Teams

As teamwork becomes ever-more important in a new age of remote work, it ...

Towards Adjustable Autonomy for the Real World

Adjustable autonomy refers to entities dynamically varying their own aut...

Evaluating and Rewarding Teamwork Using Cooperative Game Abstractions

Can we predict how well a team of individuals will perform together? How...

Commander's Intent: A Dataset and Modeling Approach for Human-AI Task Specification in Strategic Play

Effective Human-AI teaming requires the ability to communicate the goals...

Reagent: Converting Ordinary Webpages into Interactive Software Agents

We introduce Reagent, a technology that readily converts ordinary webpag...

Minimum Coverage Sets for Training Robust Ad Hoc Teamwork Agents

Robustly cooperating with unseen agents and human partners presents sign...

Please sign up or login with your details

Forgot password? Click here to reset