Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Gradient based policy optimisation. #41

Closed
patxikuku opened this issue Apr 2, 2020 · 4 comments
Closed

Gradient based policy optimisation. #41

patxikuku opened this issue Apr 2, 2020 · 4 comments

Comments

@patxikuku
Copy link

Hello,

if I understood correctly, the authors of PILCO uses a gradient based method
for optimising the policy. In the current implementation it doesn't seem to the
case, you use L-BFGS-B without giving the computation of the jacobian.

Did you make any experiments using a gradient based method ?

@nrontsis
Copy link
Owner

nrontsis commented Apr 2, 2020

Gradients are computed and used in L-BFGS-B. This is the whole point of using TensorFlow. Perhaps this is not immediately obvious when examining the code, because the gradient computation is handled via GPflow.

@nrontsis nrontsis closed this as completed Apr 2, 2020
@fuku10
Copy link

fuku10 commented May 18, 2022

Hello,
Does it mean using numerical calculated gradient, not analytically calculated gradient?

@nrontsis
Copy link
Owner

It’s neither, it’s via automatic differentiation.

@fuku10
Copy link

fuku10 commented May 19, 2022

I thought the minimize() automatically calculate the gradient using the finite-difference method.
(In case of scipy.optimize.minimize;
"If None or False, the gradient will be estimated using 2-point finite difference estimation with an absolute step size.")
https://docs.scipy.org/doc/scipy/reference/generated/scipy.optimize.minimize.html

Anyway, I'll study TensorFlow and GPflow.
Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants