toyCarIRL icon indicating copy to clipboard operation
toyCarIRL copied to clipboard

Weights for the reward function during the model learning

Open AlirezaShamsoshoara opened this issue 3 years ago • 0 comments

Hi @jangirrishabh, This repository is awesome and well-explained. I want to thank you for the great content and code. And I just have a question regarding the learning.py. My question is: before training your NN model, you defined some weights in line: https://github.com/jangirrishabh/toyCarIRL/blob/2eff036e594a787299d1e4cc82e46f0f9b21308f/learning.py#L206 and fed them into the carmunk to get the immediate reward and the new state based on the taken action to update the Y vector in the mini-batch process method. I was wondering how you defined the weights (weights for the reward function). Because later, you use this trained model in the toy_car_IRL.py to update the policy and reconstruct the weights for the reward function. So do those weights affect the trained NN model or they are just some random values?

AlirezaShamsoshoara avatar Mar 09 '21 02:03 AlirezaShamsoshoara