Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Weights for the reward function during the model learning #10

Open
AlirezaShamsoshoara opened this issue Mar 9, 2021 · 0 comments
Open

Comments

@AlirezaShamsoshoara
Copy link

Hi @jangirrishabh,
This repository is awesome and well-explained. I want to thank you for the great content and code.
And I just have a question regarding the learning.py.
My question is: before training your NN model, you defined some weights in line:

weights = [ 0.04924175 ,-0.36950358 ,-0.15510825 ,-0.65179867 , 0.2985827 , -0.23237454 , 0.21222881 ,-0.47323531]

and fed them into the carmunk to get the immediate reward and the new state based on the taken action to update the Y vector in the mini-batch process method. I was wondering how you defined the weights (weights for the reward function). Because later, you use this trained model in the toy_car_IRL.py to update the policy and reconstruct the weights for the reward function. So do those weights affect the trained NN model or they are just some random values?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant