pfrl
pfrl copied to clipboard
PFRL: a PyTorch-based deep reinforcement learning library
Noticed that [here ](https://github.com/pfnet/pfrl/blob/master/pfrl/agents/soft_actor_critic.py#L281) the `log_prob` variable is computed before the udpate of the actor while on SAC's [repo ](https://github.com/rail-berkeley/softlearning/blob/master/softlearning/algorithms/sac.py#L246) it is recomputed after the actor update (the [paper](https://arxiv.org/abs/1812.05905) also...
Gym v26 makes a number of changing to the core API, `reset` and `step` in particular, along with `render` and `seed`. Could you either pin the gym version to `
The Monitor from gym.wrappers has been deprecated. At the moment, you can;t even "import pflr", as immediately you are greeted with this: ImportError: cannot import name 'Monitor' from 'gym.wrappers'
Hi, I'm trying to setup a DQN agent with a graph attention layer. The agent can take one of 3 actions. For some reason, when I run the training function,...
Current pfrl does not support snapshot of training, which is important in many job systems such as Kubernetes. This PR support saving and loading snapshot including replay buffer. ## Done...
In Python 3.8 the default mode of multiprocessing for macOS was changed For reference: https://github.com/chainer/chainerrl/issues/572
Hindsight Experience Replay with bit-flipping example: https://arxiv.org/abs/1707.01495
Depends on https://github.com/pfnet/pfrl/pull/80. Resolves https://github.com/pfnet/pfrl/issues/6. Results: 
I have a custom environment with a [MultiDiscrete](https://github.com/openai/gym/blob/master/gym/spaces/multi_discrete.py) action space. The MultiDiscrete action space allows controlling an agent with n-dimensional discrete action spaces. In my environment, I have 4 dimensions...