This repository contains code for the paper Performative Reinforcement Learning.
The repository is structured as follows:
src/
: This folder contains all the source code files required for generating the experiments' data and figures.data/
: This folder is where all the data will be generated.figures/
: This folder is where all the figures will be generated.limiting_envs/
: This folder is for storing visualizations of the environment.
Before running the scripts, please install the following prerequisites.
Python3
matplotlib
numpy
copy
itertools
time
cvxpy
cvxopt
click
multiprocessing
statistics
json
contextlib
joblib
tqdm
os
cmath
To recreate the results of our paper, you will need to run the following scripts. Each of these scripts implements one of the methods described in the paper.
python run_experiment.py --fbeta=10
python run_experiment.py --gradient
python run_experiment.py --sampling
python run_experiment.py --gradient --sampling --etas 1
python run_experiment.py --policy_gradient
python run_experiment.py --sampling --lagrangian
After running the above scripts, new plots will be generated in the figures directory.
The following are not included in the paper:
- For the experiment repeated gradient ascent with finite samples the corresponding suboptimality gap is also computed
For any questions or comments, contact strianta@mpi-sws.org.