Interpreting systems as solving POMDPs: a step towards a formal understanding of agency

09/04/2022
by   Martin Biehl, et al.
0

Under what circumstances can a system be said to have beliefs and goals, and how do such agency-related features relate to its physical state? Recent work has proposed a notion of interpretation map, a function that maps the state of a system to a probability distribution representing its beliefs about an external world. Such a map is not completely arbitrary, as the beliefs it attributes to the system must evolve over time in a manner that is consistent with Bayes' theorem, and consequently the dynamics of a system constrain its possible interpretations. Here we build on this approach, proposing a notion of interpretation not just in terms of beliefs but in terms of goals and actions. To do this we make use of the existing theory of partially observable Markov processes (POMDPs): we say that a system can be interpreted as a solution to a POMDP if it not only admits an interpretation map describing its beliefs about the hidden state of a POMDP but also takes actions that are optimal according to its belief state. An agent is then a system together with an interpretation of this system as a POMDP solution. Although POMDPs are not the only possible formulation of what it means to have a goal, this nevertheless represents a step towards a more general formal definition of what it means for a system to be an agent.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/27/2021

Interpreting Dynamical Systems as Bayesian Reasoners

A central concept in active inference is that the internal states of a p...
research
11/06/2015

An Action Language for Multi-Agent Domains: Foundations

In multi-agent domains (MADs), an agent's action may not just change the...
research
10/06/2021

Efficient Multi-agent Epistemic Planning: Teaching Planners About Nested Belief

Many AI applications involve the interaction of multiple autonomous agen...
research
08/06/2022

Recurrent networks, hidden states and beliefs in partially observable environments

Reinforcement learning aims to learn optimal policies from interaction w...
research
09/28/2022

A Doxastic Characterisation of Autonomous Decisive Systems

A highly autonomous system (HAS) has to assess the situation it is in an...
research
04/18/2019

Asymptotic Behavior of Bayesian Learners with Misspecified Models

We consider an agent who represents uncertainty about her environment vi...
research
03/15/2012

Distribution over Beliefs for Memory Bounded Dec-POMDP Planning

We propose a new point-based method for approximate planning in Dec-POMD...

Please sign up or login with your details

Forgot password? Click here to reset