Skip to content

Repository for the publication "Learning the Optimal Power Flow: Environment Design Matters".

License

Notifications You must be signed in to change notification settings

Digitalized-Energy-Systems/rl-opf-env-design

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Learning the Optimal Power Flow: Environment Design Matters

This is the accompanying repository to the publication "Learning the Optimal Power Flow: Environment Design Matters".

Note that most source code implemented for this paper can be found in https://gitlab.com/thomaswolgast/drl (RL algorithms) and https://gitlab.com/thomaswolgast/mlopf (RL environments), which are work-in-progress and therefore continued in different repositories.

Installation

All experiments were performed with python 3.8. In an virtualenv, run pip install -r requirements.txt to install all dependencies in the right version at publication time (not the most recent).

Note: torch sometimes makes problems and needs to be installed manually before performing the previous step.

Repository structure

  • run.sh: A list of the commands performed to reproduce all experiments done for this publication. Should not be run all at once! Overall computation time will be multiple weeks. Use this file to copy-paste single commands from. Will automatically create a data/ folder with the result files.
  • LICENSE: The license used for this work (MIT).
  • requirements.txt: Reference to the two previously mentioned repositories for simple installation.
  • src/: The source code to aggregate the results and create the figures of this exact publication. The source code for running the experiments is in external repositories.
  • compact_results/: Collection of JSON files with compact representations of the results with different metrics. (created with create_results.py and used by boxplots.py)
  • results/: Optional folder that can be downloaded from Zenodo. Contains model weights, hyperparameter information, and agent performances in the course of training (e.g. MAPE every 30k steps). The model weights are required to run create_results.py. The training performances are required to plot performance during training, if required. Download this folder from Zenodo and place it in this repository to work with the data.

Attention: Sometimes, the VoltageControl environment is called QMarketEnv because of a name change during development. Both names are synonymous in the context of this work.

Contact

For questions or feedback, contact the first author Thomas Wolgast ([email protected]).

About

Repository for the publication "Learning the Optimal Power Flow: Environment Design Matters".

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published