Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

example: pytorch_lightning_mnist.py #3290

Merged
merged 1 commit into from Jan 21, 2022

Conversation

chongxiaoc
Copy link
Collaborator

Set GPU device with horovod local rank.

Signed-off-by: Chongxiao Cao chongxiaoc@uber.com

Checklist before submitting

  • Did you read the contributor guide?
  • Did you update the docs?
  • Did you write any tests to validate this change?
  • Did you update the CHANGELOG, if this change affects users?

Description

Fixes # (issue).

Review process to land

  1. All tests and other checks must succeed.
  2. At least one member of the technical steering committee must review and approve.
  3. If any member of the technical steering committee requests changes, they must be addressed.

@github-actions
Copy link

github-actions bot commented Nov 24, 2021

Unit Test Results

   394 files   -    404     394 suites   - 404   5h 48m 16s ⏱️ - 2h 55m 43s
   716 tests ±       0     575 ✔️  -      97     140 💤 +     96  1 +1 
8 573 runs   - 8 599  5 803 ✔️  - 6 319  2 769 💤  - 2 281  1 +1 

For more details on these failures, see this check.

Results for commit 0cb6ebe. ± Comparison against base commit cbce50e.

This pull request skips 96 tests.
test.parallel.test_adasum_pytorch.TorchAdasumTests ‑ test_orthogonal
test.parallel.test_adasum_pytorch.TorchAdasumTests ‑ test_parallel
test.parallel.test_mxnet1.MX1Tests ‑ test_gluon_trainer
test.parallel.test_mxnet1.MX1Tests ‑ test_gpu_required
test.parallel.test_mxnet1.MX1Tests ‑ test_horovod_allreduce_cpu_gpu_error
test.parallel.test_mxnet1.MX1Tests ‑ test_horovod_grouped_allreduce_cpu_gpu_error
test.parallel.test_mxnet2.MX2Tests ‑ test_allgather_object
test.parallel.test_mxnet2.MX2Tests ‑ test_broadcast_object
test.parallel.test_mxnet2.MX2Tests ‑ test_compression_fp16
test.parallel.test_mxnet2.MX2Tests ‑ test_gluon_trainer
…

♻️ This comment has been updated with latest results.

@github-actions
Copy link

github-actions bot commented Nov 24, 2021

Unit Test Results (with flaky tests)

   412 files   -      574     412 suites   - 574   6h 26m 56s ⏱️ - 3h 13m 27s
   716 tests ±         0     574 ✔️  -      94     140 💤 +     96  2  - 2 
8 951 runs   - 12 419  6 130 ✔️  - 8 777  2 817 💤  - 3 641  4  - 1 

For more details on these failures, see this check.

Results for commit 0cb6ebe. ± Comparison against base commit cbce50e.

This pull request skips 96 tests.
test.parallel.test_adasum_pytorch.TorchAdasumTests ‑ test_orthogonal
test.parallel.test_adasum_pytorch.TorchAdasumTests ‑ test_parallel
test.parallel.test_mxnet1.MX1Tests ‑ test_gluon_trainer
test.parallel.test_mxnet1.MX1Tests ‑ test_gpu_required
test.parallel.test_mxnet1.MX1Tests ‑ test_horovod_allreduce_cpu_gpu_error
test.parallel.test_mxnet1.MX1Tests ‑ test_horovod_grouped_allreduce_cpu_gpu_error
test.parallel.test_mxnet2.MX2Tests ‑ test_allgather_object
test.parallel.test_mxnet2.MX2Tests ‑ test_broadcast_object
test.parallel.test_mxnet2.MX2Tests ‑ test_compression_fp16
test.parallel.test_mxnet2.MX2Tests ‑ test_gluon_trainer
…

♻️ This comment has been updated with latest results.

Set GPU device with horovod local rank.

Signed-off-by: Chongxiao Cao <chongxiaoc@uber.com>
@chongxiaoc chongxiaoc merged commit 15a6aa3 into horovod:master Jan 21, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants