id doaj-96db75402bb44775a283469a9479d619
record_format Article
spelling doaj-96db75402bb44775a283469a9479d6192021-04-02T13:19:37ZengWileyIET Smart Grid2515-29472019-09-0110.1049/iet-stg.2019.0196IET-STG.2019.0196Reinforcement learning for control of flexibility providers in a residential microgridBrida V. Mbuwir0Davy Geysen1Davy Geysen2Fred Spiessens3Fred Spiessens4Geert Deconinck5VITOVITOVITOVITOVITOEnergyVilleThe smart grid paradigm and the development of smart meters have led to the availability of large volumes of data. This data is expected to assist in power system planning/operation and the transition from passive to active electricity users. With recent advances in machine learning, this data can be used to learn system dynamics. This study explores two model-free reinforcement learning (RL) techniques – policy iteration (PI) and fitted Q-iteration (FQI) for scheduling the operation of flexibility providers – battery and heat pump in a residential microgrid. The proposed algorithms are data-driven and can be easily generalised to fit the control of any flexibility provider without requiring expert knowledge to build a detailed model of the flexibility provider and/or microgrid. The algorithms are tested in multi-agent collaborative and single-agent stochastic microgrid settings – with the uncertainty due to lack of knowledge on future electricity consumption patterns and photovoltaic production. Simulation results show that PI outperforms FQI with a 7.2% increase in photovoltaic self-consumption in the multi-agent setting and a 3.7% increase in the single-agent setting. Both RL algorithms perform better than a rule-based controller, and compete with a model-based optimal controller, and are thus, a valuable alternative to model- and rule-based controllers.https://digital-library.theiet.org/content/journals/10.1049/iet-stg.2019.0196learning (artificial intelligence)multi-agent systemspower consumptiondistributed power generationphotovoltaic power systemspower engineering computingiterative methodspower generation schedulingheat pumpsstochastic processescontrol engineering computingpower generation controlresidential microgridsmart grid paradigmsmart metersmachine learningmodel-free reinforcement learning techniquessingle-agent stochastic microgrid settingsrule-based controllermodel-based optimal controllerelectricity consumption patternspower system planningrl techniquespolicy iterationpifitted q-iterationfqiheat pumpmultiagent collaborative microgrid settingsphotovoltaic production
collection DOAJ
language English
format Article
sources DOAJ
author Brida V. Mbuwir
Davy Geysen
Davy Geysen
Fred Spiessens
Fred Spiessens
Geert Deconinck
spellingShingle Brida V. Mbuwir
Davy Geysen
Davy Geysen
Fred Spiessens
Fred Spiessens
Geert Deconinck
Reinforcement learning for control of flexibility providers in a residential microgrid
IET Smart Grid
learning (artificial intelligence)
multi-agent systems
power consumption
distributed power generation
photovoltaic power systems
power engineering computing
iterative methods
power generation scheduling
heat pumps
stochastic processes
control engineering computing
power generation control
residential microgrid
smart grid paradigm
smart meters
machine learning
model-free reinforcement learning techniques
single-agent stochastic microgrid settings
rule-based controller
model-based optimal controller
electricity consumption patterns
power system planning
rl techniques
policy iteration
pi
fitted q-iteration
fqi
heat pump
multiagent collaborative microgrid settings
photovoltaic production
author_facet Brida V. Mbuwir
Davy Geysen
Davy Geysen
Fred Spiessens
Fred Spiessens
Geert Deconinck
author_sort Brida V. Mbuwir
title Reinforcement learning for control of flexibility providers in a residential microgrid
title_short Reinforcement learning for control of flexibility providers in a residential microgrid
title_full Reinforcement learning for control of flexibility providers in a residential microgrid
title_fullStr Reinforcement learning for control of flexibility providers in a residential microgrid
title_full_unstemmed Reinforcement learning for control of flexibility providers in a residential microgrid
title_sort reinforcement learning for control of flexibility providers in a residential microgrid
publisher Wiley
series IET Smart Grid
issn 2515-2947
publishDate 2019-09-01
description The smart grid paradigm and the development of smart meters have led to the availability of large volumes of data. This data is expected to assist in power system planning/operation and the transition from passive to active electricity users. With recent advances in machine learning, this data can be used to learn system dynamics. This study explores two model-free reinforcement learning (RL) techniques – policy iteration (PI) and fitted Q-iteration (FQI) for scheduling the operation of flexibility providers – battery and heat pump in a residential microgrid. The proposed algorithms are data-driven and can be easily generalised to fit the control of any flexibility provider without requiring expert knowledge to build a detailed model of the flexibility provider and/or microgrid. The algorithms are tested in multi-agent collaborative and single-agent stochastic microgrid settings – with the uncertainty due to lack of knowledge on future electricity consumption patterns and photovoltaic production. Simulation results show that PI outperforms FQI with a 7.2% increase in photovoltaic self-consumption in the multi-agent setting and a 3.7% increase in the single-agent setting. Both RL algorithms perform better than a rule-based controller, and compete with a model-based optimal controller, and are thus, a valuable alternative to model- and rule-based controllers.
topic learning (artificial intelligence)
multi-agent systems
power consumption
distributed power generation
photovoltaic power systems
power engineering computing
iterative methods
power generation scheduling
heat pumps
stochastic processes
control engineering computing
power generation control
residential microgrid
smart grid paradigm
smart meters
machine learning
model-free reinforcement learning techniques
single-agent stochastic microgrid settings
rule-based controller
model-based optimal controller
electricity consumption patterns
power system planning
rl techniques
policy iteration
pi
fitted q-iteration
fqi
heat pump
multiagent collaborative microgrid settings
photovoltaic production
url https://digital-library.theiet.org/content/journals/10.1049/iet-stg.2019.0196
work_keys_str_mv AT bridavmbuwir reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid
AT davygeysen reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid
AT davygeysen reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid
AT fredspiessens reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid
AT fredspiessens reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid
AT geertdeconinck reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid
_version_ 1721565496289001472