Skip to content

Autograd functional API should enable_grad #47543

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 3 commits into from

Conversation

iramazanli
Copy link
Contributor

@iramazanli iramazanli commented Nov 7, 2020

Fixes #44640

@dr-ci
Copy link

dr-ci bot commented Nov 7, 2020

💊 CI failures summary and remediations

As of commit 3833c39 (more details on the Dr. CI page):


None of the CI failures appear to be your fault 💚



❄️ 1 failure tentatively classified as flaky

but reruns have not yet been triggered to confirm:

See CircleCI build pytorch_xla_linux_bionic_py3_6_clang9_test (1/1)

Step: "Run tests" (full log | diagnosis details | 🔁 rerun) ❄️

Mar 11 02:42:13 RuntimeError: tensorflow/compiler/xla/xla_client/xrt_local_service.cc:56 : Check failed: tensorflow::NewServer(server_def, &server_) == ::tensorflow::Status::OK() (Unknown: Could not start gRPC server vs. OK)
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch_xla-1.9-py3.6-linux-x86_64.egg/torch_xla/distributed/xla_multiprocessing.py", line 314, in _setup_replication
Mar 11 02:42:13     device = xm.xla_device()
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch_xla-1.9-py3.6-linux-x86_64.egg/torch_xla/core/xla_model.py", line 231, in xla_device
Mar 11 02:42:13     devkind=devkind if devkind is not None else None)
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch_xla-1.9-py3.6-linux-x86_64.egg/torch_xla/core/xla_model.py", line 136, in get_xla_supported_devices
Mar 11 02:42:13     xla_devices = _DEVICES.value
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch_xla-1.9-py3.6-linux-x86_64.egg/torch_xla/utils/utils.py", line 32, in value
Mar 11 02:42:13     self._value = self._gen_fn()
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch_xla-1.9-py3.6-linux-x86_64.egg/torch_xla/core/xla_model.py", line 18, in <lambda>
Mar 11 02:42:13     _DEVICES = xu.LazyProperty(lambda: torch_xla._XLAC._xla_get_devices())
Mar 11 02:42:13 RuntimeError: tensorflow/compiler/xla/xla_client/xrt_local_service.cc:56 : Check failed: tensorflow::NewServer(server_def, &server_) == ::tensorflow::Status::OK() (Unknown: Could not start gRPC server vs. OK)
Mar 11 02:42:13 Traceback (most recent call last):
Mar 11 02:42:13   File "/var/lib/jenkins/workspace/xla/test/test_mp_collective_permute.py", line 34, in <module>
Mar 11 02:42:13     xmp.spawn(_mp_fn, args=())
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch_xla-1.9-py3.6-linux-x86_64.egg/torch_xla/distributed/xla_multiprocessing.py", line 394, in spawn
Mar 11 02:42:13     start_method=start_method)
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch/multiprocessing/spawn.py", line 188, in start_processes
Mar 11 02:42:13     while not context.join():
Mar 11 02:42:13   File "/opt/conda/lib/python3.6/site-packages/torch/multiprocessing/spawn.py", line 144, in join
Mar 11 02:42:13     exit_code=exitcode
Mar 11 02:42:13 torch.multiprocessing.spawn.ProcessExitedException: process 0 terminated with exit code 17

This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.

Please report bugs/suggestions to the (internal) Dr. CI Users group.

@iramazanli
Copy link
Contributor Author

More efficient implementation creating graphs only when its necessary

Copy link
Collaborator

@albanD albanD left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I don't think you wanted to commit the .nojekyll file right?

Copy link
Contributor Author

@iramazanli iramazanli left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yes I don't want :) so i removed the file

Copy link
Collaborator

@albanD albanD left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM
You need to fix the lint though
And you can rebase on top of the viable/strict branch to make sure you don't get any CI failures due to a broken master.

Copy link
Collaborator

@albanD albanD left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM thanks for removing the jekyll file.

@albanD
Copy link
Collaborator

albanD commented Feb 19, 2021

btw the jekyll file is back in the PR and should be removed before merging.

@codecov
Copy link

codecov bot commented Mar 9, 2021

Codecov Report

Merging #47543 (4dde815) into master (53c77e7) will decrease coverage by 0.00%.
The diff coverage is 97.10%.

@@            Coverage Diff             @@
##           master   #47543      +/-   ##
==========================================
- Coverage   77.65%   77.64%   -0.01%     
==========================================
  Files        1869     1869              
  Lines      182327   182348      +21     
==========================================
+ Hits       141583   141591       +8     
- Misses      40744    40757      +13     

Copy link
Collaborator

@albanD albanD left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Just a small update to reduce code duplication, good to go otherwise.

Thanks for the rebase.

Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@iramazanli has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@facebook-github-bot
Copy link
Contributor

@iramazanli merged this pull request in f364e49.

xsacha pushed a commit to xsacha/pytorch that referenced this pull request Mar 31, 2021
Summary:
Fixes pytorch#44640

Pull Request resolved: pytorch#47543

Reviewed By: albanD

Differential Revision: D26965136

Pulled By: iramazanli

fbshipit-source-id: 1dd46b9402bb670c0e165db684712e26c1a2036f
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

autograd functional APIs should enable_grad
3 participants