-
Notifications
You must be signed in to change notification settings - Fork 21.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Complex autograd doc fix #46258
Complex autograd doc fix #46258
Conversation
[ghstack-poisoned]
docs/source/notes/autograd.rst
Outdated
@@ -222,7 +222,7 @@ The short version: | |||
the gradients are computed under the assumption that the function is a part of a larger real-valued | |||
loss function :math:`g(input)=L`. The gradient computed is :math:`\frac{\partial L}{\partial z^*}` | |||
(note the conjugation of z), which is precisely the direction of the step | |||
you should take in gradient descent. Thus, all the existing optimizers work out of | |||
you should take in gradient ascent. Thus, all the existing optimizers work out of |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
While it is true, I don't think anyone is doing gradient ascent anymore :D
I think that not specifying is better here: "which is precisely the direction expected by gradient-based optimization algorithms." What do you think?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yeah that sounds good! alternately we can also consider:
" ... (note the conjugation of z), the negative of which is precisely the direction of steepest descent in Gradient Descent algorithm." it might be useful to make reference to Gradient Descent since (as you mentioned) practically majority of people use Gradient Descent as opposed to other alternatives that fall under "gradient-based optimization algorithms" ...
💊 CI failures summary and remediationsAs of commit 512668c (more details on the Dr. CI page): 💚 💚 Looks good so far! There are no failures yet. 💚 💚 This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.Please report bugs/suggestions on the GitHub issue tracker or post in the (internal) Dr. CI Users group. This comment has been revised 8 times. |
[ghstack-poisoned]
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
docs/source/notes/autograd.rst
Outdated
(note the conjugation of z), which is precisely the direction of the step | ||
you should take in gradient descent. Thus, all the existing optimizers work out of | ||
(note the conjugation of z), the negative of which is precisely the direction of steepest descent | ||
in Gradient Descent algorithm.. Thus, all the existing optimizers work out of |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: missing a "used"
[ghstack-poisoned]
ghstack-source-id: 03ba913f8b673163b5c4e9456c7eff3529eb237b Pull Request resolved: #46258
(note the conjugation of z), which is precisely the direction of the step | ||
you should take in gradient descent. Thus, all the existing optimizers work out of | ||
(note the conjugation of z), the negative of which is precisely the direction of steepest descent | ||
used in Gradient Descent algorithm.. Thus, all the existing optimizers work out of |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: two periods
@anjali411 merged this pull request in ac245f6. |
Stack from ghstack:
Differential Revision: D24286512