Reinforcement learning for control of flexibility providers in a residential microgrid
The smart grid paradigm and the development of smart meters have led to the availability of large volumes of data. This data is expected to assist in power system planning/operation and the transition from passive to active electricity users. With recent advances in machine learning, this data can b...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2019-09-01
|
Series: | IET Smart Grid |
Subjects: | |
Online Access: | https://digital-library.theiet.org/content/journals/10.1049/iet-stg.2019.0196 |
id |
doaj-96db75402bb44775a283469a9479d619 |
---|---|
record_format |
Article |
spelling |
doaj-96db75402bb44775a283469a9479d6192021-04-02T13:19:37ZengWileyIET Smart Grid2515-29472019-09-0110.1049/iet-stg.2019.0196IET-STG.2019.0196Reinforcement learning for control of flexibility providers in a residential microgridBrida V. Mbuwir0Davy Geysen1Davy Geysen2Fred Spiessens3Fred Spiessens4Geert Deconinck5VITOVITOVITOVITOVITOEnergyVilleThe smart grid paradigm and the development of smart meters have led to the availability of large volumes of data. This data is expected to assist in power system planning/operation and the transition from passive to active electricity users. With recent advances in machine learning, this data can be used to learn system dynamics. This study explores two model-free reinforcement learning (RL) techniques – policy iteration (PI) and fitted Q-iteration (FQI) for scheduling the operation of flexibility providers – battery and heat pump in a residential microgrid. The proposed algorithms are data-driven and can be easily generalised to fit the control of any flexibility provider without requiring expert knowledge to build a detailed model of the flexibility provider and/or microgrid. The algorithms are tested in multi-agent collaborative and single-agent stochastic microgrid settings – with the uncertainty due to lack of knowledge on future electricity consumption patterns and photovoltaic production. Simulation results show that PI outperforms FQI with a 7.2% increase in photovoltaic self-consumption in the multi-agent setting and a 3.7% increase in the single-agent setting. Both RL algorithms perform better than a rule-based controller, and compete with a model-based optimal controller, and are thus, a valuable alternative to model- and rule-based controllers.https://digital-library.theiet.org/content/journals/10.1049/iet-stg.2019.0196learning (artificial intelligence)multi-agent systemspower consumptiondistributed power generationphotovoltaic power systemspower engineering computingiterative methodspower generation schedulingheat pumpsstochastic processescontrol engineering computingpower generation controlresidential microgridsmart grid paradigmsmart metersmachine learningmodel-free reinforcement learning techniquessingle-agent stochastic microgrid settingsrule-based controllermodel-based optimal controllerelectricity consumption patternspower system planningrl techniquespolicy iterationpifitted q-iterationfqiheat pumpmultiagent collaborative microgrid settingsphotovoltaic production |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Brida V. Mbuwir Davy Geysen Davy Geysen Fred Spiessens Fred Spiessens Geert Deconinck |
spellingShingle |
Brida V. Mbuwir Davy Geysen Davy Geysen Fred Spiessens Fred Spiessens Geert Deconinck Reinforcement learning for control of flexibility providers in a residential microgrid IET Smart Grid learning (artificial intelligence) multi-agent systems power consumption distributed power generation photovoltaic power systems power engineering computing iterative methods power generation scheduling heat pumps stochastic processes control engineering computing power generation control residential microgrid smart grid paradigm smart meters machine learning model-free reinforcement learning techniques single-agent stochastic microgrid settings rule-based controller model-based optimal controller electricity consumption patterns power system planning rl techniques policy iteration pi fitted q-iteration fqi heat pump multiagent collaborative microgrid settings photovoltaic production |
author_facet |
Brida V. Mbuwir Davy Geysen Davy Geysen Fred Spiessens Fred Spiessens Geert Deconinck |
author_sort |
Brida V. Mbuwir |
title |
Reinforcement learning for control of flexibility providers in a residential microgrid |
title_short |
Reinforcement learning for control of flexibility providers in a residential microgrid |
title_full |
Reinforcement learning for control of flexibility providers in a residential microgrid |
title_fullStr |
Reinforcement learning for control of flexibility providers in a residential microgrid |
title_full_unstemmed |
Reinforcement learning for control of flexibility providers in a residential microgrid |
title_sort |
reinforcement learning for control of flexibility providers in a residential microgrid |
publisher |
Wiley |
series |
IET Smart Grid |
issn |
2515-2947 |
publishDate |
2019-09-01 |
description |
The smart grid paradigm and the development of smart meters have led to the availability of large volumes of data. This data is expected to assist in power system planning/operation and the transition from passive to active electricity users. With recent advances in machine learning, this data can be used to learn system dynamics. This study explores two model-free reinforcement learning (RL) techniques – policy iteration (PI) and fitted Q-iteration (FQI) for scheduling the operation of flexibility providers – battery and heat pump in a residential microgrid. The proposed algorithms are data-driven and can be easily generalised to fit the control of any flexibility provider without requiring expert knowledge to build a detailed model of the flexibility provider and/or microgrid. The algorithms are tested in multi-agent collaborative and single-agent stochastic microgrid settings – with the uncertainty due to lack of knowledge on future electricity consumption patterns and photovoltaic production. Simulation results show that PI outperforms FQI with a 7.2% increase in photovoltaic self-consumption in the multi-agent setting and a 3.7% increase in the single-agent setting. Both RL algorithms perform better than a rule-based controller, and compete with a model-based optimal controller, and are thus, a valuable alternative to model- and rule-based controllers. |
topic |
learning (artificial intelligence) multi-agent systems power consumption distributed power generation photovoltaic power systems power engineering computing iterative methods power generation scheduling heat pumps stochastic processes control engineering computing power generation control residential microgrid smart grid paradigm smart meters machine learning model-free reinforcement learning techniques single-agent stochastic microgrid settings rule-based controller model-based optimal controller electricity consumption patterns power system planning rl techniques policy iteration pi fitted q-iteration fqi heat pump multiagent collaborative microgrid settings photovoltaic production |
url |
https://digital-library.theiet.org/content/journals/10.1049/iet-stg.2019.0196 |
work_keys_str_mv |
AT bridavmbuwir reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid AT davygeysen reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid AT davygeysen reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid AT fredspiessens reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid AT fredspiessens reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid AT geertdeconinck reinforcementlearningforcontrolofflexibilityprovidersinaresidentialmicrogrid |
_version_ |
1721565496289001472 |