Approximate policy iteration using neural networks for storage problems

10/04/2019
by   Trivikram Dokka, et al.
0

We consider the stochastic single node energy storage problem (SNES) and revisit Approximate Policy Iteration (API) to solve SNES. We show that the performance of API can be boosted by using neural networks as an approximation architecture at the policy evaluation stage. To achieve this, we use a model different to that in literature with aggregate variables reducing the dimensionality of the decision vector, which in turn makes it viable to use neural network predictions in the policy improvement stage. We show that performance improvement by neural networks is even more significant in the case when charging efficiency of storage systems is low.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/12/2014

Approximate Policy Iteration Schemes: A Comparison

We consider the infinite-horizon discounted optimal control problem form...
research
04/12/2010

Dynamic Policy Programming

In this paper, we propose a novel policy iteration method, called dynami...
research
07/17/2021

Greedification Operators for Policy Optimization: Investigating Forward and Reverse KL Divergences

Approximate Policy Iteration (API) algorithms alternate between (approxi...
research
10/27/2022

Confident Approximate Policy Iteration for Efficient Local Planning in q^π-realizable MDPs

We consider approximate dynamic programming in γ-discounted Markov decis...
research
11/09/2020

Multiagent Rollout and Policy Iteration for POMDP with Application to Multi-Robot Repair Problems

In this paper we consider infinite horizon discounted dynamic programmin...
research
03/02/2023

Dynamic discretization discovery under hard node storage constraints

The recently developed dynamic discretization discovery (DDD) is a power...

Please sign up or login with your details

Forgot password? Click here to reset