From db6ef7f585d47005208bfa9d145c4442b99e13b5 Mon Sep 17 00:00:00 2001 From: Ryan Strauss Date: Fri, 13 Dec 2024 20:36:22 -0800 Subject: [PATCH] Update README.md --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 90eb895..e7cc6ff 100644 --- a/README.md +++ b/README.md @@ -21,13 +21,13 @@ because the entire loop of environment simulation, agent interaction, and policy XLA program and executed on CPUs, GPUs, or TPUs. More specifically, the implementations in Dopamax follow the Anakin Podracer architecture -- see [this paper][2] for more details. -> [!WARNING] -> Note that this repository is not actively maintained and is subject to breaking changes at any time. - ## Supported Algorithms - [Proximal Policy Optimization (PPO)](dopamax/agents/anakin/ppo.py) - [Deep Q-Network (DQN)](dopamax/agents/anakin/dqn.py) +- [Deep Deterministic Policy Gradients (DDPG)](dopamax/agents/anakin/ddpg.py) +- [Twin Delayed DDPG (TD3)](dopamax/agents/anakin/ddpg.py) +- [Soft Actor Critic](dopamax/agents/anakin/sac.py) - [AlphaZero](dopamax/agents/anakin/alphazero.py) ## Installation