New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Adadelta optimizer #644
Conversation
missing the GPU impl, did you forget to git add the file? cause the code references it but the file is missing ; ) also, I think a few lines are over 80 chars. but otherwise, LGTM. |
@zffchen78: when you get a chance can you take a look at this? |
Can one of the admins verify this patch? |
LGTM |
@vrv ok I will add the cuda part and do the edits ! |
Two suggestions:
I think there is a mistake: |
FYI, there's a couple of bugs, namely the += shoulda been = my impl is here (GPU included + separate lr included + sparse): |
Hi, mistobaan, please let us know how you plan to proceed on this? I can see the following work items:
1 & 2 must be settled before we can add this op since it affects the op's interface / specification and will be hard to fix later.
You can get 1&2 done and merged first; and collaborate w/ others to get 3&4 done or you can incrementally get them done later. |
Hi! On Saturday, January 16, 2016, zffchen78 notifications@github.com wrote:
LinkedIn: http://linkedin.com/in/fmilo Github: http://github.com/Mistobaan/Simplicity, consistency, and repetition - that's how you get through. (Jack |
18dfed6
to
16a7c89
Compare
@Mistobaan: any updates? |
@vrv I think the operation themselves are up to date and correct. I was not able to create a solid testing for GPU as I don't have a supported GPU environment at the moment. I just rebased the patch. |
Ok, we'll probably soon have GPU testing, right @martinwicke ? :). Then we can try pulling this in. |
@tensorflow-jenkins: test this please |
Can one of the admins verify this patch? |
tensorflow/core/kernels/training_ops.cc:338:8: error: template-id 'operator()<>' for 'void tensorflow::functor::ApplyAdadelta<Eigen::GpuDevice, double>::operator()(const GPUDevice&, tensorflow::TTypes<double, 1, long int>::Flat, tensorflow::TTypes<double, 1, long int>::Flat, tensorflow::TTypes<double, 1, long int>::Flat, tensorflow::TTypes<double, 1, long int>::ConstScalar, tensorflow::TTypes<double, 1, long int>::ConstScalar, tensorflow::TTypes<double, 1, long int>::ConstScalar, tensorflow::TTypes<double, 1, long int>::ConstFlat)' does not match any template declaration |
(ping this thread when this is ready -- it looks like you added a commit but don't know if it's ready) |
@Mistobaan, any update on this? I'm also really interested in trying AdaDelta. Thanks! |
# limitations under the License. | ||
# ============================================================================== | ||
|
||
"""Tests for Momentum.""" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Docstring references Momentum instead of AdaDelta.
7753631
to
ad01e6f
Compare
@tensorflow-jenkins: test this please |
@Mistobaan: I think this looks good pending the GPU tests finishing -- the only other thing I think we need is for you to run bazel-bin/tensorflow/core/ops/compat/update_ops tensorflow/core/ops and add that updated file to your commit for tracking backwards compatibility. |
@vrv I run and added the modified files as you requested. |
@tensorflow-jenkins: test this please |
Looks like nothing built -- can you double check and verify that your commit compiles and passes in at least one config? |
@vrv let's try again I ran the updated command on an old branch and force pushed |
@tensorflow-jenkins: test this please |
…axPotentialBlockSize enable hipOccupancyMaxPotentialBlockSize
This is my attempt at #516 . Looking for some early feedback as is my first attempt.