pfrl icon indicating copy to clipboard operation
pfrl copied to clipboard

PFRL: a PyTorch-based deep reinforcement learning library

Results 41 pfrl issues
Sort by recently updated
recently updated
newest added

Noticed that [here ](https://github.com/pfnet/pfrl/blob/master/pfrl/agents/soft_actor_critic.py#L281) the `log_prob` variable is computed before the udpate of the actor while on SAC's [repo ](https://github.com/rail-berkeley/softlearning/blob/master/softlearning/algorithms/sac.py#L246) it is recomputed after the actor update (the [paper](https://arxiv.org/abs/1812.05905) also...

Gym v26 makes a number of changing to the core API, `reset` and `step` in particular, along with `render` and `seed`. Could you either pin the gym version to `

The Monitor from gym.wrappers has been deprecated. At the moment, you can;t even "import pflr", as immediately you are greeted with this: ImportError: cannot import name 'Monitor' from 'gym.wrappers'

Hi, I'm trying to setup a DQN agent with a graph attention layer. The agent can take one of 3 actions. For some reason, when I run the training function,...

Current pfrl does not support snapshot of training, which is important in many job systems such as Kubernetes. This PR support saving and loading snapshot including replay buffer. ## Done...

In Python 3.8 the default mode of multiprocessing for macOS was changed For reference: https://github.com/chainer/chainerrl/issues/572

Hindsight Experience Replay with bit-flipping example: https://arxiv.org/abs/1707.01495

Depends on https://github.com/pfnet/pfrl/pull/80. Resolves https://github.com/pfnet/pfrl/issues/6. Results: ![her_bit_flip_dqn](https://user-images.githubusercontent.com/10005453/97736859-a0c2f700-1b1f-11eb-85ce-5fcf8e69d4dd.png)

I have a custom environment with a [MultiDiscrete](https://github.com/openai/gym/blob/master/gym/spaces/multi_discrete.py) action space. The MultiDiscrete action space allows controlling an agent with n-dimensional discrete action spaces. In my environment, I have 4 dimensions...