DeepAI AI Chat
Log In Sign Up

Optimal Demand Response Using Device Based Reinforcement Learning

01/08/2014
by   Zheng Wen, et al.
Stanford University
0

Demand response (DR) for residential and small commercial buildings is estimated to account for as much as 65 of DR, and previous work shows that a fully automated Energy Management System (EMS) is a necessary prerequisite to DR in these areas. In this paper, we propose a novel EMS formulation for DR problems in these sectors. Specifically, we formulate a fully automated EMS's rescheduling problem as a reinforcement learning (RL) problem, and argue that this RL problem can be approximately solved by decomposing it over device clusters. Compared with existing formulations, our new formulation (1) does not require explicitly modeling the user's dissatisfaction on job rescheduling, (2) enables the EMS to self-initiate jobs, (3) allows the user to initiate more flexible requests and (4) has a computational complexity linear in the number of devices. We also demonstrate the simulation results of applying Q-learning, one of the most popular and classical RL algorithms, to a representative example.

READ FULL TEXT

page 1

page 2

page 3

page 4

02/10/2018

More Robust Doubly Robust Off-policy Evaluation

We study the problem of off-policy evaluation (OPE) in reinforcement lea...
12/18/2022

Stochastic Economic Dispatch Considering Demand Response and Endogenous Uncertainty

This paper considers endogenous uncertainty (EnU) in the stochastic econ...
04/29/2019

A supervised-learning-based strategy for optimal demand response of an HVAC System

The large thermal capacity of buildings enables heating, ventilating, an...