DeepRL-TensorFlow2
DeepRL-TensorFlow2 copied to clipboard
🐋 Simple implementations of various popular Deep Reinforcement Learning algorithms using TensorFlow2
[No changes have been made to the code. tensorflow version is 2.2, will this affect it? 
'Viewer' object has no attribute 'isopen' File "E:\anaconda\envs\tf2\lib\site-packages\gym\envs\classic_control\rendering.py", line 81, in close AttributeError: 'Viewer' object has no attribute 'isopen' Traceback (most recent call last): File "E:\anaconda\envs\tf2\lib\site-packages\gym\envs\classic_control\rendering.py", line 165, in __del__...
Hello everyone. I am trying to use A3C continuous. But I am getting some error saying "unrecognized arguments". Please see the attached picture.   How to solve this?
https://github.com/marload/DeepRL-TensorFlow2/blob/876266d9a5fcf7d8a7c7e3afd8b110085b32b615/PPO/PPO_Discrete.py#L151-L154 https://github.com/marload/DeepRL-TensorFlow2/blob/876266d9a5fcf7d8a7c7e3afd8b110085b32b615/PPO/PPO_Continuous.py#L167-L170 In `PPO_Discrete` each reward is multiplied by `0.01` and in `PPO_Continuous` reward is also modified. I don't understand why do these modification, what does these modification do?
In the Actor net, It seems that from_logit should be set to False in tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True) as you added a softmax in the last layer:)