Loss Bounds for Approximate Influence-Based Abstraction

11/03/2020
by   Elena Congeduti, et al.
0

Sequential decision making techniques hold great promise to improve the performance of many real-world systems, but computational complexity hampers their principled application. Influence-based abstraction aims to gain leverage by modeling local subproblems together with the 'influence' that the rest of the system exerts on them. While computing exact representations of such influence might be intractable, learning approximate representations offers a promising approach to enable scalable solutions. This paper investigates the performance of such approaches from a theoretical perspective. The primary contribution is the derivation of sufficient conditions on approximate influence representations that can guarantee solutions with small value loss. In particular we show that neural networks trained with cross entropy are well suited to learn approximate influence representations. Moreover, we provide a sample based formulation of the bounds, which reduces the gap to applications. Finally, driven by our theoretical insights, we propose approximation error estimators, which empirically reveal to correlate well with the value loss.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/22/2019

A Sufficient Statistic for Influence in Structured Multiagent Environments

Making decisions in complex environments is a key challenge in artificia...
research
02/18/2015

Influence-Optimistic Local Values for Multiagent Planning --- Extended Version

Recent years have seen the development of methods for multiagent plannin...
research
01/21/2023

Improving Deep Regression with Ordinal Entropy

In computer vision, it is often observed that formulating regression pro...
research
02/09/2021

Enhancing Audio Augmentation Methods with Consistency Learning

Data augmentation is an inexpensive way to increase training data divers...
research
01/15/2017

Near Optimal Behavior via Approximate State Abstraction

The combinatorial explosion that plagues planning and reinforcement lear...
research
04/16/2021

Lower Bounds on Cross-Entropy Loss in the Presence of Test-time Adversaries

Understanding the fundamental limits of robust supervised learning has e...
research
04/09/2018

Policy Gradient With Value Function Approximation For Collective Multiagent Planning

Decentralized (PO)MDPs provide an expressive framework for sequential de...

Please sign up or login with your details

Forgot password? Click here to reset