Skip to content

Conversation

@minitu
Copy link
Contributor

@minitu minitu commented Sep 29, 2023

This PR adds support for hysteresis in AMP gradient scale update.

Jaemin Choi and others added 5 commits September 20, 2023 15:56
…s by tricking torch autograd (NVIDIA#1715)

* input grad checks out

* adding clamp gamma

* Both old and proposed implementation checks out

* 2 tests not yet passed due to numerical issues

* mem_eff works

* fast-layer-norm done

* Moving mem-eff to templates

* Relax tolerance for memory efficient backward

* Fix backward api of python
* Add distopt support for param syncs with non-floating-point dtypes

Signed-off-by: Tim Moon <tmoon@nvidia.com>

* Update apex/contrib/optimizers/distributed_fused_adam.py

Co-authored-by: Masaki Kozuki <mkozuki@nvidia.com>

---------

Signed-off-by: Tim Moon <tmoon@nvidia.com>
Co-authored-by: Masaki Kozuki <mkozuki@nvidia.com>
@minitu
Copy link
Contributor Author

minitu commented Sep 30, 2023

@crcrpar Thanks, addressed your comments

@crcrpar crcrpar merged commit 6a77872 into NVIDIA:master Sep 30, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants