rl-baselines3-zoo
rl-baselines3-zoo copied to clipboard
A training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included.
## Description ## Motivation and Context - [ ] I have raised an issue to propose this change ([required](https://github.com/DLR-RM/stable-baselines3/blob/master/CONTRIBUTING.md) for new features and bug fixes) ## Types of changes -...
It would be nice to have a cli that is available everywhere (with completion too), something like: ```bash sb3 train --algo ppo --env CartPole-v1 sb3 train --algo ppo --env CartPole-v1...
### Question Is it possible to pass parameters to a validation environment? ### Additional context I'm using rl-zoo to tune a DDPG agent on a custom environment. From the command...
Right now, only hyperparmeters that are searched by default can have their params dict be copied and reused due to naming issues. This should be extended to hyperparameters that are...
Running the make type script (for PR #140) results in an error I am not able to decipher that does not seem likely to be my fault: ``` justinterry@preacherMan2 rl-baselines3-zoo...
Building off the jsons of the best hyperparameters saved in https://github.com/DLR-RM/rl-baselines3-zoo/pull/140, I have a script that takes each and runs trains it for 10 times (by default) to determine which...
In my trials tuning both pettingzoo gym environments, all optuna jobs on one node have just simply stopped twice now. There have been no system errors, nothing printed to stdout,...
Enabling the possibility to run --optimize with the --trained-agent flag would be great ! In my case, I pre-trained an agent on a simplified task and want to continue training...
I needed to do a run with PPO on a Gym environment on my cluster to make sure everything is working right before moving onto tuning PettingZoo environments, so I...
See https://optuna.readthedocs.io/en/latest/reference/generated/optuna.study.Study.html#optuna.study.Study