Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Problems with the exact gradient method in original paper #2

Open
V-Enzo opened this issue May 27, 2020 · 0 comments
Open

Problems with the exact gradient method in original paper #2

V-Enzo opened this issue May 27, 2020 · 0 comments

Comments

@V-Enzo
Copy link

V-Enzo commented May 27, 2020

Hi xiangyu, sorry to interrupt you. I recently want to implement the exact gradient method (gradient descent for LQR with gradient oracle), which is proved in lemma 1. P_k could be calculated iteratively with P = Q, while \sigma_K is hard to compute, as it sums all xx^T from t=0 to infinity. To program it, I set the infinite length to 1000. and run for 20000 epochs with learning rate, state_dim, action_dim equals to 1e-3, 100, 20, separately . However, the result doesn't seem to converge as shown on page 38 of the original paper. Since I am new and working alone on this problem, I really appreciate you can give me some insights! Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant