DRL-FlappyBird
DRL-FlappyBird copied to clipboard
Playing Flappy Bird Using Deep Reinforcement Learning (Based on Deep Q Learning DQN using Tensorflow)
Hi, Thanks for your nice code and documentation. I saw the report from Kevin Chen [http://cs229.stanford.edu/proj2015/362_report.pdf] where he experimented with three difficulty levels (easy, medium, hard) of the game. Can...
I have no idea about the meaning of copyTargetQNetwork. Why did we need QValueT to eval the QValue_batch? In order to let training process more stable ?
I read from [here](https://github.com/songrotek/DRL-FlappyBird/blob/master/BrainDQN_NIPS.py#L89-L108). Why do the program only use the current state and the next state? Why only using the two state can work? Thank you @songrotek
I find from [here ](https://github.com/songrotek/DRL-FlappyBird/blob/master/BrainDQN_NIPS.py#L88) that all the rewards are add into the deque. We need to sample the 1 and -1 reward from the deque to use them. So...