Seeing the Forest Despite the Trees: Large Scale Spatial-Temporal Decision Making

05/09/2012
by   Mark Crowley, et al.
0

We introduce a challenging real-world planning problem where actions must be taken at each location in a spatial area at each point in time. We use forestry planning as the motivating application. In Large Scale Spatial-Temporal (LSST) planning problems, the state and action spaces are defined as the cross-products of many local state and action spaces spread over a large spatial area such as a city or forest. These problems possess state uncertainty, have complex utility functions involving spatial constraints and we generally must rely on simulations rather than an explicit transition model. We define LSST problems as reinforcement learning problems and present a solution using policy gradients. We compare two different policy formulations: an explicit policy that identifies each location in space and the action to take there; and an abstract policy that defines the proportion of actions to take across all locations in space. We show that the abstract policy is more robust and achieves higher rewards with far fewer parameters than the elementary policy. This abstract policy is also a better fit to the properties that practitioners in LSST problem domains require for such methods to be widely useful.

READ FULL TEXT
research
06/05/2019

Lifelong Learning with a Changing Action Set

In many real-world sequential decision making problems, the number of av...
research
04/13/2021

Learning and Planning in Complex Action Spaces

Many important real-world problems have action spaces that are high-dime...
research
07/02/2019

Procedure Planning in Instructional Videos

We propose a new challenging task: procedure planning in instructional v...
research
05/20/2017

Learning to Factor Policies and Action-Value Functions: Factored Action Space Representations for Deep Reinforcement learning

Deep Reinforcement Learning (DRL) methods have performed well in an incr...
research
04/04/2014

Scalable Planning and Learning for Multiagent POMDPs: Extended Version

Online, sample-based planning algorithms for POMDPs have shown great pro...
research
10/28/2021

Equivariant Q Learning in Spatial Action Spaces

Recently, a variety of new equivariant neural network model architecture...
research
06/10/2020

Marginal Utility for Planning in Continuous or Large Discrete Action Spaces

Sample-based planning is a powerful family of algorithms for generating ...

Please sign up or login with your details

Forgot password? Click here to reset