In this repository we're trying to solve 2 openAI Gym's env using two policy based methods: Hill climbing and Cross-Entropy method
Hill_Climbing.ipynb
is an implementation of hill climbing with adaptive noise scaling for OpenAI Gym's Cartpole environment.
CEM.ipynb
is an implementation of the cross-entropy method for OpenAI Gym's MountainCarContinuous environment.
For more well explained methods for policy based method here's a good blog:
http://kvfrans.com/simple-algoritms-for-solving-cartpole/
--> corresponding github: https://github.com/kvfrans/openai-cartpole