Skip to content

Commit

Permalink
conditionally import amp_C (#1211)
Browse files Browse the repository at this point in the history
  • Loading branch information
crcrpar authored Nov 10, 2021
1 parent 2205cff commit 25bfcb9
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion apex/transformer/pipeline_parallel/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -20,10 +20,11 @@
from torch.nn.parallel import DistributedDataParallel

from apex.multi_tensor_apply import multi_tensor_applier
import amp_C
from apex.transformer import parallel_state
from apex.transformer.microbatches import build_num_microbatches_calculator
from apex.transformer.pipeline_parallel._timers import _Timers
if multi_tensor_applier.available:
import amp_C


_GLOBAL_ARGS = None
Expand Down

0 comments on commit 25bfcb9

Please sign in to comment.