Asynchronous-Methods-for-Deep-Reinforcement-Learning
Asynchronous-Methods-for-Deep-Reinforcement-Learning copied to clipboard
Using a paper from Google DeepMind I've developed a new version of the DQN using threads exploration instead of memory replay as explain in here: http://arxiv.org/pdf/1602.01783v1.pdf I used the one-...
when i start the thread it show this exception Exception in thread Thread-31: Traceback (most recent call last): File "/home/anderson/.conda/envs/tensorflow/lib/python3.6/threading.py", line 916, in _bootstrap_inner self.run() File "/home/anderson/.conda/envs/tensorflow/lib/python3.6/threading.py", line 864, in...
The paper states that the final epsilons should be [0.1, 0.01, 0.5]. But I noticed in your code they are [0.01, 0.01, 0.05] (Strangely there are two 0.01s). Is this...
Hi @Zeta36 Great project! I'm trying to run some experiments with the code. It seems that currently the code uses threading with tensorflow, and from my observation, the training loop...
Hi zeta: I saw your reply and came here to have a look. This is so cool! I really need a method that can avoid memory replay since the memory...
Hello, In the [asynchronous dqn paper](http://arxiv.org/pdf/1602.01783v1.pdf), they also described an on policy method, the advantage actor-critic (A3C), which achieved better results than others, do you currently have any plan to...