Abc11c

Results 9 issues of Abc11c

Is there an oracle similar to [BabyAIBot](https://github.com/Farama-Foundation/Minigrid/blob/master/tests/test_baby_ai_bot.py) for the minigrid environment to get some demos to train a rl agent ? Thanks!

### Question Is there an oracle similar to [BabyAIBot](https://github.com/Farama-Foundation/Minigrid/blob/master/tests/test_baby_ai_bot.py) for the minigrid environment to get some demos to train a rl agent ? Or generate human demonstrations ? Thanks!

### Question Hi, I was wondering if it's possible to add more than one goal or introduce sub-goals in the maze_2d environments ? I'm currently looking at `set_target` func [here](https://github.com/Farama-Foundation/D4RL/blob/71a9549f2091accff93eeff68f1f3ab2c0e0a288/d4rl/pointmaze/maze_model.py#L211)...

### Model/Pipeline/Scheduler description Hi, I was looking into the paper, [D3PM](https://openreview.net/forum?id=h7-XixPCAL) was wondering which scheduler to choose from particularly my problem deals with discrete action space and will be using...

stale

Hi @xuchen-ethz Thanks for releasing the code! Was wondering if the filter is capable of handling background points? I understand it primarily filters diverged and converged points based on the...

Hi @cvignac , I'm unclear about the sampling part in the code base. Like the `add_noise` step [here](https://github.com/cvignac/DiGress/blob/e88372a837adcac88d24387f3c01d04ae2d36012/src/diffusion_model_discrete.py#L110) I'm wondering if there is a `prev_sample` while performing reverse process? ```...

Hi @alex-petrenko, Thank you for such a great code! I'm working on some RL algorithms (intrinsic reward-based ones for the MiniGrid environment) based on the version [here](https://github.com/kakaobrain/leco). I was wondering...

Hi @thomasschmied, Thank you for releasing the code! I want to quickly benchmark some of the metaworld v2 environments but stuck at mujuco setup I have the following doubts and...

Hi @jayLEE0301, Thank you for releasing the code! I wanted to quickly benchmark the non-goal conditioned ant env, but the training feels slow, am I missing something ? Is it...