Empowerment for Continuous Agent-Environment Systems

01/31/2012
by   Tobias Jung, et al.
0

This paper develops generalizations of empowerment to continuous states. Empowerment is a recently introduced information-theoretic quantity motivated by hypotheses about the efficiency of the sensorimotor loop in biological organisms, but also from considerations stemming from curiosity-driven learning. Empowemerment measures, for agent-environment systems with stochastic transitions, how much influence an agent has on its environment, but only that influence that can be sensed by the agent sensors. It is an information-theoretic generalization of joint controllability (influence on environment) and observability (measurement by sensors) of the environment by the agent, both controllability and observability being usually defined in control theory as the dimensionality of the control/observation spaces. Earlier work has shown that empowerment has various interesting and relevant properties, e.g., it allows us to identify salient states using only the dynamics, and it can act as intrinsic reward without requiring an external reward. However, in this previous work empowerment was limited to the case of small-scale and discrete domains and furthermore state transition probabilities were assumed to be known. The goal of this paper is to extend empowerment to the significantly more important and relevant case of continuous vector-valued state spaces and initially unknown state transition probabilities. The continuous state space is addressed by Monte-Carlo approximation; the unknown transitions are addressed by model learning and prediction for which we apply Gaussian processes regression with iterated forecasting. In a number of well-known continuous control tasks we examine the dynamics induced by empowerment and include an application to exploration and online model learning.

READ FULL TEXT
research
10/12/2019

Influence-Based Multi-Agent Exploration

Intrinsically motivated reinforcement learning aims to address the explo...
research
06/22/2020

Information Theoretic Regret Bounds for Online Nonlinear Control

This work studies the problem of sequential control in an unknown, nonli...
research
12/04/2019

Learning Efficient Representation for Intrinsic Motivation

Mutual Information between agent Actions and environment States (MIAS) q...
research
01/17/2022

Summarising and Comparing Agent Dynamics with Contrastive Spatiotemporal Abstraction

We introduce a data-driven, model-agnostic technique for generating a hu...
research
05/22/2019

The Journey is the Reward: Unsupervised Learning of Influential Trajectories

Unsupervised exploration and representation learning become increasingly...
research
12/29/2022

Intrinsic Motivation in Dynamical Control Systems

Biological systems often choose actions without an explicit reward signa...
research
09/28/2015

Efficient Empowerment

Empowerment quantifies the influence an agent has on its environment. Th...

Please sign up or login with your details

Forgot password? Click here to reset