cups-rl icon indicating copy to clipboard operation
cups-rl copied to clipboard

Regarding to policy/model/weights

Open zyzhang1130 opened this issue 5 years ago • 4 comments

Do you mind to clarify is the policy/model/weights saved after each epoch/iteration? If not how should I make it happen? If yes where is it saved at? I see you just call the model 'args.model_path' in agent.py and save(self, path, filename) without specifically assigning a path or name.

Thank you for replying.

zyzhang1130 avatar Feb 03 '20 06:02 zyzhang1130

Hi, I found this piece of info quite relevant! https://github.com/Kaixhin/Rainbow/pull/58 However, I am not proficient to incorporate it into the rainbow code of yours (judging based on the time it was added, your current version rainbow probably does not have this feature). Do you mind to take a look at it?

Thank you so much

zyzhang1130 avatar Feb 03 '20 08:02 zyzhang1130

Actually the contributor of Rainbow says these lines save the model weights: https://github.com/Kaixhin/Rainbow/blob/d3afb5ad570137d675d6c7c903c050c8a19db084/main.py#L179-L181 How do you think I should incorporate it into cups-lr Rainbow code?

Thank you.

zyzhang1130 avatar Feb 03 '20 13:02 zyzhang1130

Line 79 in rainbow/test.py which is called periodically from the main.py file is the code which saves the model:

# Save model parameters if improved
if avg_reward > best_avg_reward:
    best_avg_reward = avg_reward
    dqn.save(path='weights', filename='rainbow_{}.pt'.format(num_steps))

This is therefore saved in the rainbow directory from which is related to #17. Set your evaluation-interval lower. If it still doesn't save the model after args.evaluation_interval steps let me know.

beduffy avatar Feb 11 '20 00:02 beduffy

Sorry, can I check does the current repo allow resume training feature? Thank you.

zyzhang1130 avatar Feb 20 '20 03:02 zyzhang1130