
2024
- Lehna, Malte, Clara Holzhüter, Sven Tomforde, and Christoph Scholz. “HUGO – Highlighting Unseen Grid Options: Combining Deep Reinforcement Learning With a Heuristic Target Topology Approach”. Sustainable Energy, Grids and Networks 39 (September 2024): 101510. doi:10.1016/j.segan.2024.101510.
@article{Lehna_2024,
author = {Lehna, Malte and Holzhüter, Clara and Tomforde, Sven and Scholz, Christoph},
journal = {Sustainable Energy, Grids and Networks},
keywords = {rl4ces},
month = {09},
pages = 101510,
publisher = {Elsevier BV},
title = {HUGO – Highlighting Unseen Grid Options: Combining deep reinforcement learning with a heuristic target topology approach},
volume = 39,
year = 2024
}%0 Journal Article
%1 Lehna_2024
%A Lehna, Malte
%A Holzhüter, Clara
%A Tomforde, Sven
%A Scholz, Christoph
%D 2024
%I Elsevier BV
%J Sustainable Energy, Grids and Networks
%P 101510
%R 10.1016/j.segan.2024.101510
%T HUGO – Highlighting Unseen Grid Options: Combining deep reinforcement learning with a heuristic target topology approach
%U http://dx.doi.org/10.1016/j.segan.2024.101510
%V 39 - Lehna, Malte, Mohamed Hassouna, Dmitry Degtyar, Sven Tomforde, and Christoph Scholz. “Fault Detection for Agents in Power Grid Topology Optimization: A Comprehensive Analysis”. In Machine Learning for Sustainable Power Systems (ML4SPS), ECML. Springer, 2024.Optimizing the topology of transmission networks using Deep Reinforcement Learning (DRL) has increasingly come into focus. Various DRL agents have been proposed, which are mostly benchmarked on the Grid2Op environment from the Learning to Run a Power Network (L2RPN) challenges. The environments have many advantages with their realistic chronics and underlying power flow backends. However, the interpretation of agent survival or failure is not always clear, as there are a variety of potential causes. In this work, we focus on the failures of the power grid simulation to identify patterns and detect them in advance. We collect the failed chronics of three different agents on the WCCI 2022 L2RPN environment, totaling about 40k data points. By clustering, we are able to detect five distinct clusters, identifying common failure types. Further, we propose a multi-class prediction approach to detect failures beforehand and evaluate five different prediction models. Here, the Light Gradient-Boosting Machine (LightGBM) shows the best failure detection performance, with an accuracy of 86%. It also correctly identifies in 91% of the time failure and survival observations. Finally, we provide a detailed feature importance analysis that identifies critical features and regions in the grid.
@inproceedings{lehna2024fault,
abstract = {Optimizing the topology of transmission networks using Deep Reinforcement Learning (DRL) has increasingly come into focus. Various DRL agents have been proposed, which are mostly benchmarked on the Grid2Op environment from the Learning to Run a Power Network (L2RPN) challenges. The environments have many advantages with their realistic chronics and underlying power flow backends. However, the interpretation of agent survival or failure is not always clear, as there are a variety of potential causes. In this work, we focus on the failures of the power grid simulation to identify patterns and detect them in advance. We collect the failed chronics of three different agents on the WCCI 2022 L2RPN environment, totaling about 40k data points. By clustering, we are able to detect five distinct clusters, identifying common failure types. Further, we propose a multi-class prediction approach to detect failures beforehand and evaluate five different prediction models. Here, the Light Gradient-Boosting Machine (LightGBM) shows the best failure detection performance, with an accuracy of 86%. It also correctly identifies in 91% of the time failure and survival observations. Finally, we provide a detailed feature importance analysis that identifies critical features and regions in the grid.},
author = {Lehna, Malte and Hassouna, Mohamed and Degtyar, Dmitry and Tomforde, Sven and Scholz, Christoph},
booktitle = {Machine Learning for Sustainable Power Systems (ML4SPS), ECML},
keywords = {rl4ces},
note = {(accepted)},
publisher = {Springer},
title = {Fault Detection for Agents in Power Grid Topology Optimization: A Comprehensive Analysis},
year = 2024
}%0 Conference Paper
%1 lehna2024fault
%A Lehna, Malte
%A Hassouna, Mohamed
%A Degtyar, Dmitry
%A Tomforde, Sven
%A Scholz, Christoph
%B Machine Learning for Sustainable Power Systems (ML4SPS), ECML
%D 2024
%I Springer
%T Fault Detection for Agents in Power Grid Topology Optimization: A Comprehensive Analysis
%X Optimizing the topology of transmission networks using Deep Reinforcement Learning (DRL) has increasingly come into focus. Various DRL agents have been proposed, which are mostly benchmarked on the Grid2Op environment from the Learning to Run a Power Network (L2RPN) challenges. The environments have many advantages with their realistic chronics and underlying power flow backends. However, the interpretation of agent survival or failure is not always clear, as there are a variety of potential causes. In this work, we focus on the failures of the power grid simulation to identify patterns and detect them in advance. We collect the failed chronics of three different agents on the WCCI 2022 L2RPN environment, totaling about 40k data points. By clustering, we are able to detect five distinct clusters, identifying common failure types. Further, we propose a multi-class prediction approach to detect failures beforehand and evaluate five different prediction models. Here, the Light Gradient-Boosting Machine (LightGBM) shows the best failure detection performance, with an accuracy of 86%. It also correctly identifies in 91% of the time failure and survival observations. Finally, we provide a detailed feature importance analysis that identifies critical features and regions in the grid. - Hassouna, Mohamed, Clara Holzhüter, Pawel Lytaev, Josephine M. Thomas, Bernhard Sick, and Christoph Scholz. “Graph Reinforcement Learning in Power Grids: A Survey.”. CoRR abs/2407.04522 (2024). http://dblp.uni-trier.de/db/journals/corr/corr2407.html#abs-2407-04522.
@article{journals/corr/abs-2407-04522,
author = {Hassouna, Mohamed and Holzhüter, Clara and Lytaev, Pawel and Thomas, Josephine M. and Sick, Bernhard and Scholz, Christoph},
journal = {CoRR},
keywords = {rl4ces},
title = {Graph Reinforcement Learning in Power Grids: A Survey.},
volume = {abs/2407.04522},
year = 2024
}%0 Journal Article
%1 journals/corr/abs-2407-04522
%A Hassouna, Mohamed
%A Holzhüter, Clara
%A Lytaev, Pawel
%A Thomas, Josephine M.
%A Sick, Bernhard
%A Scholz, Christoph
%D 2024
%J CoRR
%T Graph Reinforcement Learning in Power Grids: A Survey.
%U http://dblp.uni-trier.de/db/journals/corr/corr2407.html#abs-2407-04522
%V abs/2407.04522 - Lehna, Malte, Clara Holzhüter, Sven Tomforde, and Christoph Scholz. “HUGO -- Highlighting Unseen Grid Options: Combining Deep Reinforcement Learning With a Heuristic Target Topology Approach”. doi:https://doi.org/10.1016/j.segan.2024.101510.
@misc{lehna2024hugohighlightingunseen,
author = {Lehna, Malte and Holzhüter, Clara and Tomforde, Sven and Scholz, Christoph},
keywords = {rl4ces},
title = {HUGO -- Highlighting Unseen Grid Options: Combining Deep Reinforcement Learning with a Heuristic Target Topology Approach},
year = 2024
}%0 Generic
%1 lehna2024hugohighlightingunseen
%A Lehna, Malte
%A Holzhüter, Clara
%A Tomforde, Sven
%A Scholz, Christoph
%D 2024
%R https://doi.org/10.1016/j.segan.2024.101510
%T HUGO -- Highlighting Unseen Grid Options: Combining Deep Reinforcement Learning with a Heuristic Target Topology Approach
%U https://arxiv.org/abs/2405.00629
2023
- Lehna, Malte, Jan Viebahn, Antoine Marot, Sven Tomforde, and Christoph Scholz. “Managing Power Grids through Topology Actions: A Comparative Study Between Advanced Rule-Based and Reinforcement Learning Agents”. Energy and AI 14 (2023): 100276. doi:10.1016/j.egyai.2023.100276.The operation of electricity grids has become increasingly complex due to the current upheaval and the increase in renewable energy production. As a consequence, active grid management is reaching its limits with conventional approaches. In the context of the Learning to Run a Power Network (L2RPN) challenge, it has been shown that Reinforcement Learning (RL) is an efficient and reliable approach with considerable potential for automatic grid operation. In this article, we analyse the submitted agent from Binbinchen and provide novel strategies to improve the agent, both for the RL and the rule-based approach. The main improvement is a N-1 strategy, where we consider topology actions that keep the grid stable, even if one line is disconnected. More, we also propose a topology reversion to the original grid, which proved to be beneficial. The improvements are tested against reference approaches on the challenge test sets and are able to increase the performance of the rule-based agent by 27%. In direct comparison between rule-based and RL agent we find similar performance. However, the RL agent has a clear computational advantage. We also analyse the behaviour in an exemplary case in more detail to provide additional insights. Here, we observe that through the N-1 strategy, the actions of both the rule-based and the RL agent become more diversified.
@article{lehna2023managing,
abstract = {The operation of electricity grids has become increasingly complex due to the current upheaval and the increase in renewable energy production. As a consequence, active grid management is reaching its limits with conventional approaches. In the context of the Learning to Run a Power Network (L2RPN) challenge, it has been shown that Reinforcement Learning (RL) is an efficient and reliable approach with considerable potential for automatic grid operation. In this article, we analyse the submitted agent from Binbinchen and provide novel strategies to improve the agent, both for the RL and the rule-based approach. The main improvement is a N-1 strategy, where we consider topology actions that keep the grid stable, even if one line is disconnected. More, we also propose a topology reversion to the original grid, which proved to be beneficial. The improvements are tested against reference approaches on the challenge test sets and are able to increase the performance of the rule-based agent by 27%. In direct comparison between rule-based and RL agent we find similar performance. However, the RL agent has a clear computational advantage. We also analyse the behaviour in an exemplary case in more detail to provide additional insights. Here, we observe that through the N-1 strategy, the actions of both the rule-based and the RL agent become more diversified.},
author = {Lehna, Malte and Viebahn, Jan and Marot, Antoine and Tomforde, Sven and Scholz, Christoph},
journal = {Energy and AI},
keywords = {rl4ces},
pages = 100276,
publisher = {Energy and AI},
title = {Managing power grids through topology actions: A comparative study between advanced rule-based and reinforcement learning agents},
volume = 14,
year = 2023
}%0 Journal Article
%1 lehna2023managing
%A Lehna, Malte
%A Viebahn, Jan
%A Marot, Antoine
%A Tomforde, Sven
%A Scholz, Christoph
%D 2023
%I Energy and AI
%J Energy and AI
%P 100276
%R 10.1016/j.egyai.2023.100276
%T Managing power grids through topology actions: A comparative study between advanced rule-based and reinforcement learning agents
%U https://www.sciencedirect.com/science/article/pii/S2666546823000484?via%3Dihub
%V 14
%X The operation of electricity grids has become increasingly complex due to the current upheaval and the increase in renewable energy production. As a consequence, active grid management is reaching its limits with conventional approaches. In the context of the Learning to Run a Power Network (L2RPN) challenge, it has been shown that Reinforcement Learning (RL) is an efficient and reliable approach with considerable potential for automatic grid operation. In this article, we analyse the submitted agent from Binbinchen and provide novel strategies to improve the agent, both for the RL and the rule-based approach. The main improvement is a N-1 strategy, where we consider topology actions that keep the grid stable, even if one line is disconnected. More, we also propose a topology reversion to the original grid, which proved to be beneficial. The improvements are tested against reference approaches on the challenge test sets and are able to increase the performance of the rule-based agent by 27%. In direct comparison between rule-based and RL agent we find similar performance. However, the RL agent has a clear computational advantage. We also analyse the behaviour in an exemplary case in more detail to provide additional insights. Here, we observe that through the N-1 strategy, the actions of both the rule-based and the RL agent become more diversified. - Heinrich, René, Christoph Scholz, Stephan Vogt, and Malte Lehna. “Targeted Adversarial Attacks on Wind Power Forecasts”. Machine Learning 113, no. 2 (2023): 863–889. doi:10.1007/s10994-023-06396-9.In recent years, researchers proposed a variety of deep learning models for wind power forecasting. These models predict the wind power generation of wind farms or entire regions more accurately than traditional machine learning algorithms or physical models. However, latest research has shown that deep learning models can often be manipulated by adversarial attacks. Since wind power forecasts are essential for the stability of modern power systems, it is important to protect them from this threat. In this work, we investigate the vulnerability of two different forecasting models to targeted, semi-targeted, and untargeted adversarial attacks. We consider a Long Short-Term Memory (LSTM) network for predicting the power generation of individual wind farms and a Convolutional Neural Network (CNN) for forecasting the wind power generation throughout Germany. Moreover, we propose the Total Adversarial Robustness Score (TARS), an evaluation metric for quantifying the robustness of regression models to targeted and semi-targeted adversarial attacks. It assesses the impact of attacks on the model's performance, as well as the extent to which the attacker's goal was achieved, by assigning a score between 0 (very vulnerable) and 1 (very robust). In our experiments, the LSTM forecasting model was fairly robust and achieved a TARS value of over 0.78 for all adversarial attacks investigated. The CNN forecasting model only achieved TARS values below 0.10 when trained ordinarily, and was thus very vulnerable. Yet, its robustness could be significantly improved by adversarial training, which always resulted in a TARS above 0.46.
@article{heinrich2023targeted,
abstract = {In recent years, researchers proposed a variety of deep learning models for wind power forecasting. These models predict the wind power generation of wind farms or entire regions more accurately than traditional machine learning algorithms or physical models. However, latest research has shown that deep learning models can often be manipulated by adversarial attacks. Since wind power forecasts are essential for the stability of modern power systems, it is important to protect them from this threat. In this work, we investigate the vulnerability of two different forecasting models to targeted, semi-targeted, and untargeted adversarial attacks. We consider a Long Short-Term Memory (LSTM) network for predicting the power generation of individual wind farms and a Convolutional Neural Network (CNN) for forecasting the wind power generation throughout Germany. Moreover, we propose the Total Adversarial Robustness Score (TARS), an evaluation metric for quantifying the robustness of regression models to targeted and semi-targeted adversarial attacks. It assesses the impact of attacks on the model's performance, as well as the extent to which the attacker's goal was achieved, by assigning a score between 0 (very vulnerable) and 1 (very robust). In our experiments, the LSTM forecasting model was fairly robust and achieved a TARS value of over 0.78 for all adversarial attacks investigated. The CNN forecasting model only achieved TARS values below 0.10 when trained ordinarily, and was thus very vulnerable. Yet, its robustness could be significantly improved by adversarial training, which always resulted in a TARS above 0.46.},
author = {Heinrich, René and Scholz, Christoph and Vogt, Stephan and Lehna, Malte},
journal = {Machine Learning},
keywords = {rl4ces},
number = 2,
pages = {863--889},
publisher = {Springer},
title = {Targeted Adversarial Attacks on Wind Power Forecasts},
volume = 113,
year = 2023
}%0 Journal Article
%1 heinrich2023targeted
%A Heinrich, René
%A Scholz, Christoph
%A Vogt, Stephan
%A Lehna, Malte
%D 2023
%I Springer
%J Machine Learning
%N 2
%P 863--889
%R 10.1007/s10994-023-06396-9
%T Targeted Adversarial Attacks on Wind Power Forecasts
%V 113
%X In recent years, researchers proposed a variety of deep learning models for wind power forecasting. These models predict the wind power generation of wind farms or entire regions more accurately than traditional machine learning algorithms or physical models. However, latest research has shown that deep learning models can often be manipulated by adversarial attacks. Since wind power forecasts are essential for the stability of modern power systems, it is important to protect them from this threat. In this work, we investigate the vulnerability of two different forecasting models to targeted, semi-targeted, and untargeted adversarial attacks. We consider a Long Short-Term Memory (LSTM) network for predicting the power generation of individual wind farms and a Convolutional Neural Network (CNN) for forecasting the wind power generation throughout Germany. Moreover, we propose the Total Adversarial Robustness Score (TARS), an evaluation metric for quantifying the robustness of regression models to targeted and semi-targeted adversarial attacks. It assesses the impact of attacks on the model's performance, as well as the extent to which the attacker's goal was achieved, by assigning a score between 0 (very vulnerable) and 1 (very robust). In our experiments, the LSTM forecasting model was fairly robust and achieved a TARS value of over 0.78 for all adversarial attacks investigated. The CNN forecasting model only achieved TARS values below 0.10 when trained ordinarily, and was thus very vulnerable. Yet, its robustness could be significantly improved by adversarial training, which always resulted in a TARS above 0.46.