Skip to content

Commit

Permalink
Remove redundant stmts after pytest parametrize
Browse files Browse the repository at this point in the history
  • Loading branch information
sadra-barikbin committed Apr 20, 2022
1 parent 48ea731 commit ee21110
Showing 1 changed file with 1 addition and 43 deletions.
44 changes: 1 addition & 43 deletions tests/ignite/engine/test_create_supervised.py
Original file line number Diff line number Diff line change
Expand Up @@ -148,7 +148,7 @@ def _test_create_supervised_trainer_have_grad_after_iteration(

@trainer.on(Events.ITERATION_COMPLETED)
def _():
assert model.weight.data.grad != 0
assert model.weight.grad != 0

trainer.run(data)

Expand Down Expand Up @@ -373,7 +373,6 @@ def _test_create_evaluation_step(
def test_create_supervised_trainer(gradient_accumulation_steps):
_test_create_supervised_trainer_wrong_accumulation()
_test_create_supervised_trainer(gradient_accumulation_steps=gradient_accumulation_steps)
_test_create_supervised_trainer(gradient_accumulation_steps=gradient_accumulation_steps)
_test_create_mocked_supervised_trainer()
_test_create_supervised_trainer_have_grad_after_iteration(gradient_accumulation_steps=gradient_accumulation_steps)

Expand All @@ -382,7 +381,6 @@ def test_create_supervised_trainer(gradient_accumulation_steps):
def test_create_supervised_trainer_with_cpu(gradient_accumulation_steps):
_test_create_supervised_trainer_wrong_accumulation(trainer_device="cpu")
_test_create_supervised_trainer(gradient_accumulation_steps=gradient_accumulation_steps, trainer_device="cpu")
_test_create_supervised_trainer(gradient_accumulation_steps=gradient_accumulation_steps, trainer_device="cpu")
_test_create_mocked_supervised_trainer(trainer_device="cpu")
_test_create_supervised_trainer_have_grad_after_iteration(
trainer_device="cpu", gradient_accumulation_steps=gradient_accumulation_steps
Expand All @@ -395,9 +393,6 @@ def test_create_supervised_trainer_traced_with_cpu(gradient_accumulation_steps):
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps, trainer_device="cpu", trace=True
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps, trainer_device="cpu", trace=True
)
_test_create_mocked_supervised_trainer(trainer_device="cpu", trace=True)
_test_create_supervised_trainer_have_grad_after_iteration(
trainer_device="cpu", trace=True, gradient_accumulation_steps=gradient_accumulation_steps
Expand Down Expand Up @@ -458,11 +453,6 @@ def test_create_supervised_trainer_on_cuda(gradient_accumulation_steps):
model_device=model_device,
trainer_device=trainer_device,
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps,
model_device=model_device,
trainer_device=trainer_device,
)
_test_create_mocked_supervised_trainer(model_device=model_device, trainer_device=trainer_device)
_test_create_supervised_trainer_have_grad_after_iteration(
model_device=model_device,
Expand All @@ -485,12 +475,6 @@ def test_create_supervised_trainer_on_cuda_amp(gradient_accumulation_steps):
trainer_device=trainer_device,
amp_mode="amp",
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps,
model_device=model_device,
trainer_device=trainer_device,
amp_mode="amp",
)
_test_create_mocked_supervised_trainer(model_device=model_device, trainer_device=trainer_device, amp_mode="amp")
_test_create_supervised_trainer_have_grad_after_iteration(
model_device=model_device,
Expand All @@ -515,13 +499,6 @@ def test_create_supervised_trainer_on_cuda_amp_scaler(gradient_accumulation_step
amp_mode="amp",
scaler=True,
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps,
model_device=model_device,
trainer_device=trainer_device,
amp_mode="amp",
scaler=True,
)
_test_create_mocked_supervised_trainer(
model_device=model_device, trainer_device=trainer_device, amp_mode="amp", scaler=True
)
Expand All @@ -533,13 +510,6 @@ def test_create_supervised_trainer_on_cuda_amp_scaler(gradient_accumulation_step
amp_mode="amp",
scaler=scaler,
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps,
model_device=model_device,
trainer_device=trainer_device,
amp_mode="amp",
scaler=scaler,
)
_test_create_mocked_supervised_trainer(
model_device=model_device, trainer_device=trainer_device, amp_mode="amp", scaler=scaler
)
Expand Down Expand Up @@ -567,12 +537,6 @@ def test_create_supervised_trainer_on_cuda_apex(gradient_accumulation_steps):
trainer_device=trainer_device,
amp_mode="apex",
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps,
model_device=model_device,
trainer_device=trainer_device,
amp_mode="apex",
)
_test_create_mocked_supervised_trainer(model_device=model_device, trainer_device=trainer_device, amp_mode="apex")
_test_create_supervised_trainer_have_grad_after_iteration(
model_device=model_device,
Expand Down Expand Up @@ -608,11 +572,6 @@ def test_create_supervised_trainer_on_tpu(gradient_accumulation_steps):
model_device=model_device,
trainer_device=trainer_device,
)
_test_create_supervised_trainer(
gradient_accumulation_steps=gradient_accumulation_steps,
model_device=model_device,
trainer_device=trainer_device,
)
_test_create_mocked_supervised_trainer(model_device=model_device, trainer_device=trainer_device)
_test_create_supervised_trainer_have_grad_after_iteration(
model_device=model_device,
Expand All @@ -634,7 +593,6 @@ def test_create_supervised_trainer_on_tpu_amp():
def test_create_supervised_trainer_on_cuda_with_model_on_cpu(gradient_accumulation_steps):
_test_create_supervised_trainer_wrong_accumulation(trainer_device="cuda")
_test_create_supervised_trainer(gradient_accumulation_steps=gradient_accumulation_steps, trainer_device="cuda")
_test_create_supervised_trainer(gradient_accumulation_steps=gradient_accumulation_steps, trainer_device="cuda")
_test_create_mocked_supervised_trainer(trainer_device="cuda")
_test_create_supervised_trainer_have_grad_after_iteration(
trainer_device="cuda", gradient_accumulation_steps=gradient_accumulation_steps
Expand Down

0 comments on commit ee21110

Please sign in to comment.