RLeXplore is a unified, highly-modularized and plug-and-play toolkit that currently provides high-quality and reliable implementations of eight representative intrinsic reward algorithms. It used to be challenging to compare intrinsic reward algorithms due to various confounding factors, including distinct implementations, optimization strategies, and evaluation methodologies. Therefore, RLeXplore is designed to provide unified and standardized procedures for constructing, computing, and optimizing intrinsic reward modules.
The workflow of RLeXplore is illustrated as follows:
- with pip
recommended
Open a terminal and install rllte with pip
:
conda create -n rllte python=3.8
pip install rllte-core
- with git
Open a terminal and clone the repository from GitHub with git
:
git clone https://github.com/RLE-Foundation/rllte.git
pip install -e .
Now you can invoke the intrinsic reward module by:
from rllte.xplore.reward import ICM, RIDE, ...
Type | Modules |
---|---|
Count-based | PseudoCounts, RND, E3B |
Curiosity-driven | ICM, Disagreement, RIDE |
Memory-based | NGU |
Information theory-based | RE3 |
Click the following links to get the code notebook:
- Quick Start
- RLeXplore with RLLTE
- RLeXplore with Stable-Baselines3
- RLeXplore with CleanRL
- Exploring Mixed Intrinsic Rewards
- Custom Intrinsic Rewards
RLLTE's PPO+RLeXplore
on SuperMarioBros:
CleanRL's PPO+RLeXplore's RND
on Montezuma's Revenge:
To cite this repository in publications:
@article{yuan_roger2024rlexplore,
title={RLeXplore: Accelerating Research in Intrinsically-Motivated Reinforcement Learning},
author={Yuan, Mingqi and Castanyer, Roger Creus and Li, Bo and Jin, Xin and Berseth, Glen and Zeng, Wenjun},
journal={arXiv preprint arXiv:2405.19548},
year={2024}
}