Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Avoid cloning gradient tensor in embedding backward pass (#2526)
Summary: Pull Request resolved: #2526 I found memory spike during embedding kernel backward `split_embedding_backward_codegen_rowwise_adagrad_unweghted_exact_cuda`, which was traced into the below code making a clone of the gradient tensor. This logic didn't seem to be there in the original code: https://github.com/pytorch/FBGEMM/pull/2347/files#diff-944ab49dcbcf54826cc3e1eab5e3c0c787b5a195f602c2d3052adae14c506d78. Reviewed By: ezyang Differential Revision: D56420646 fbshipit-source-id: a4e3fd6952cdaa4f1a3339980151f5dc1ce6c436
- Loading branch information