sheeprl icon indicating copy to clipboard operation
sheeprl copied to clipboard

Distributed Reinforcement Learning accelerated by Lightning Fabric

Results 27 sheeprl issues
Sort by recently updated
recently updated
newest added

Hi, when I run (DreamerV3) experiments, especially ones with a replay_ratio > 1.0, training takes quite a long time. During these runs, my hardware resources are not being used much...

I was also about to open an issue regarding Dreamer on feature-vector based (partially observable) environments where no Cnn is needed (and as a matter of fact, to also handle...

enhancement
help wanted

updates: - [github.com/PyCQA/autoflake: v2.2.1 → v2.3.1](https://github.com/PyCQA/autoflake/compare/v2.2.1...v2.3.1) - [github.com/psf/black: 23.12.1 → 24.8.0](https://github.com/psf/black/compare/23.12.1...24.8.0) - [github.com/astral-sh/ruff-pre-commit: v0.1.11 → v0.6.8](https://github.com/astral-sh/ruff-pre-commit/compare/v0.1.11...v0.6.8)

I can't find a button to reopen the issue, regarding #272, I have two questions! 1) shouldn't initial `stochastic_state` be flattened s.t. ``` stochastic_state = player.stochastic_state.view(1, 1, -1).clone() ``` in...

Hi! Sharing slight change in Dreamer V3 according to their updated(2024/04/17) manuscript https://arxiv.org/pdf/2301.04104 Also their codes are updated few hours ago https://github.com/danijar/dreamerv3 It includes change in the optimizer (LaProp), experiments...

enhancement
help wanted

Hi, I am trying to write a simple gym wrapper for an existing env. During testing, I am not facing the following issue: ``` File "/home/drt/miniconda3/envs/sheeprl/lib/python3.10/site-packages/sheeprl/algos/dreamer_v3/dreamer_v3.py", line 647, in main...

wontfix

hi, i tried out this project and it is one of the few that actually works off the shelf, thank you for your work. Is there a way to enable...

It would be nice to add a test for the environment wrappers we have. Maybe, using the dummy envs, we could do a couple of iterations and check that the...

Hi everyone, As a professional who has worked with a few RL frameworks in the past, I can confidently say that this is one of the cleanest, most user-friendly, and...

Hi, Working on an Atari environment wrapper with action input buffer with `len=N` that I want to feed as input to `mlp_keys`. Algo config: ```yaml algo: mlp_keys: encoder: [actions] ```...

enhancement