async-rl icon indicating copy to clipboard operation
async-rl copied to clipboard

Replicating "Asynchronous Methods for Deep Reinforcement Learning" (http://arxiv.org/abs/1602.01783)

Results 19 async-rl issues
Sort by recently updated
recently updated
newest added

In: python demo_a3c_ale.py [--use-lstm] I've been looking at what to put in for quite a while but I have just found it's related to ALE. I've tried putting either 'breakout'...

v_loss += (v - R) ** 2 / 2 But the original paper just calculate the derivative of the (V-R)^2 right?

I trained the model for 3000000 iterations and saved the model as "3000000.h5". But when I try to evaluate using demo_a3c_ale.py, it turns out there is an error there saying...

As titled. I found the console just print out scores too fast, where can we find the score VS training iteration records?

Hi, I just noticed: https://github.com/muupan/async-rl/blob/master/ale.py#L115 each training action is taken 4x times to the game environment? e.g. user pressed 'down' once, but in your simulated training the environment to take...

Hey, so the color transform that you use is incorrect (for example in Seaquest it sometimes causes the fishes to disappear). https://github.com/muupan/async-rl/blob/12dac595b2ad9d99c19e82e92adb825359a3a3eb/ale.py#L67-L68 You can get the correct one from the...

When I try to run the saved model as : ``` bash python demo_a3c_ale.py ../roms/breakout.bin trained_model/breakout_ff/80000000_finish.h5 ``` I get an error : ``` bash ImportError: No module named 'ale_python_interface' ```...

Hi there - I forked your code to work on [Super Mario Bros](https://github.com/ehrenbrav/async-rl/tree/eb) :) I'm using a [Nintendo emulator that I modified](https://github.com/ehrenbrav/FCEUX_Learning_Environment) to allow for programmatic control by the agent...