Model-predictive control and reinforcement learning in multi-energy system case studies

Glenn Ceusters*, Román Cantú Rodríguez, Alberte Bouso García, Rüdiger Franke, Geert Deconinck, Lieve Helsen, Ann Nowé, Maarten Messagie, Luis Ramirez Camargo

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Model predictive control (MPC) offers an optimal control technique to establish and ensure that the total operation cost of multi-energy systems remains at a minimum while fulfilling all system constraints. However, this method presumes an adequate model of the underlying system dynamics, which is prone to modelling errors and is not necessarily adaptive. This has an associated initial and ongoing project-specific engineering cost. In this paper, we present an on- and off-policy multi-objective reinforcement learning (RL) approach that does not assume a model a priori, benchmarking this against a linear MPC (LMPC — to reflect current practice, though non-linear MPC performs better) - both derived from the general optimal control problem, highlighting their differences and similarities. In a simple multi-energy system (MES) configuration case study, we show that a twin delayed deep deterministic policy gradient (TD3) RL agent offers the potential to match and outperform the perfect foresight LMPC benchmark (101.5%). This while the realistic LMPC, i.e. imperfect predictions, only achieves 98%. While in a more complex MES system configuration, the RL agent's performance is generally lower (94.6%), yet still better than the realistic LMPC (88.9%). In both case studies, the RL agents outperformed the realistic LMPC after a training period of 2 years using quarterly interactions with the environment. We conclude that reinforcement learning is a viable optimal control technique for multi-energy systems given adequate constraint handling and pre-training, to avoid unsafe interactions and long training periods, as is proposed in fundamental future work.

Original languageEnglish
Article number117634
Pages (from-to)1-12
JournalApplied Energy
Volume303
DOIs
Publication statusPublished - 1 Dec 2021
Externally publishedYes

Bibliographical note

Funding Information:
This work has been supported in part by ABB n.v. and Flemish Agency for Innovation and Entrepreneurship (VLAIO) grant HBC.2019.2613 and grant HBC.2018.0529. We also acknowledge Flanders Make for the support to the MOBI research group.

Funding Information:
This work has been supported in part by ABB n.v. and Flemish Agency for Innovation and Entrepreneurship (VLAIO) grant HBC.2019.2613 and grant HBC.2018.0529 . We also acknowledge Flanders Make for the support to the MOBI research group.

Publisher Copyright:
© 2021 Elsevier Ltd

Funding

This work has been supported in part by ABB n.v. and Flemish Agency for Innovation and Entrepreneurship (VLAIO) grant HBC.2019.2613 and grant HBC.2018.0529. We also acknowledge Flanders Make for the support to the MOBI research group. This work has been supported in part by ABB n.v. and Flemish Agency for Innovation and Entrepreneurship (VLAIO) grant HBC.2019.2613 and grant HBC.2018.0529 . We also acknowledge Flanders Make for the support to the MOBI research group.

Keywords

  • Model-predictive control
  • Multi-energy systems
  • Optimal control
  • Reinforcement learning

Fingerprint

Dive into the research topics of 'Model-predictive control and reinforcement learning in multi-energy system case studies'. Together they form a unique fingerprint.

Cite this