async-rl
async-rl copied to clipboard
Replicating "Asynchronous Methods for Deep Reinforcement Learning" (http://arxiv.org/abs/1602.01783)
In: python demo_a3c_ale.py [--use-lstm] I've been looking at what to put in for quite a while but I have just found it's related to ALE. I've tried putting either 'breakout'...
v_loss += (v - R) ** 2 / 2 But the original paper just calculate the derivative of the (V-R)^2 right?
I trained the model for 3000000 iterations and saved the model as "3000000.h5". But when I try to evaluate using demo_a3c_ale.py, it turns out there is an error there saying...
As titled. I found the console just print out scores too fast, where can we find the score VS training iteration records?
Hi, I just noticed: https://github.com/muupan/async-rl/blob/master/ale.py#L115 each training action is taken 4x times to the game environment? e.g. user pressed 'down' once, but in your simulated training the environment to take...
Hey, so the color transform that you use is incorrect (for example in Seaquest it sometimes causes the fishes to disappear). https://github.com/muupan/async-rl/blob/12dac595b2ad9d99c19e82e92adb825359a3a3eb/ale.py#L67-L68 You can get the correct one from the...
When I try to run the saved model as : ``` bash python demo_a3c_ale.py ../roms/breakout.bin trained_model/breakout_ff/80000000_finish.h5 ``` I get an error : ``` bash ImportError: No module named 'ale_python_interface' ```...
Hi there - I forked your code to work on [Super Mario Bros](https://github.com/ehrenbrav/async-rl/tree/eb) :) I'm using a [Nintendo emulator that I modified](https://github.com/ehrenbrav/FCEUX_Learning_Environment) to allow for programmatic control by the agent...