Denys Makoviichuk
Denys Makoviichuk
It looks like half done and I didn't find KL loss parts. Only policy loss.
I see you have two options for the exploration loss: symmetric_kl_with_uniform_prior and entropy. And you used symmetric KL as the loss in the vizdoom config. Is there any paper or...
zero_rnn_on_done by default is True.
It was requested in this ticket: https://github.com/Denys88/rl_games/issues/194 Two things should be done: 1) env_info returned from the vec_env should have 'env_masks' : True 2) method get_env_masks(self) should be implemented
- Added SHAC implementation - Aligned SHAC with PPO and SAC implementations. - Exposed min_lr for linear and adaptive schedules and max_lr for adaptive
Simply doesn't work. Not going to merge right now. To try vmpo one can run: python runner.py --train --file rl_games/configs/vmpo_lunar_discrete_torch.yaml
Hi, I tried to use your product but got a lot of small issues and found some lack of functionality. 1) Please don't expect that git address is something line...
We used https://gymnasium.farama.org/content/migration-guide/ and shimmy library. https://github.com/Denys88/rl_games/issues/261
https://arxiv.org/pdf/2212.06343.pdf
auto lambdaWithReturn = [](){return 5; }; std::function f = lambdaWithReturn;