Skip to content

Commit

Permalink
Linting fixes
Browse files Browse the repository at this point in the history
Signed-off-by: Alp Dener <adener@nvidia.com>
  • Loading branch information
denera committed May 22, 2024
1 parent e900ba6 commit b260d50
Showing 1 changed file with 3 additions and 5 deletions.
8 changes: 3 additions & 5 deletions transformer_engine/pytorch/distributed.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,7 @@
"""Methods needed for distributed training (DP/TP)."""
import warnings
from contextlib import contextmanager, AbstractContextManager, ContextDecorator
from typing import Any, Dict, Union, Optional, Callable, Tuple, List, final
from abc import ABC
from collections.abc import Iterable
from functools import partial
from typing import Any, Dict, Union, Optional, Callable, Tuple, List

import torch
from torch.cuda import _lazy_call, _lazy_init
Expand Down Expand Up @@ -849,7 +846,8 @@ def _fsdp_scatter_tensors(
target = t._data if isinstance(t, Float8Tensor) else t
shapes.append(target.data.shape)
safely_set_viewless_tensor_data(
target, split_tensor_into_1d_equal_chunks(target.data, fsdp_group, new_buffer=True)
target, split_tensor_into_1d_equal_chunks(
target.data, fsdp_group, new_buffer=True)
)
else:
shapes.append(None)
Expand Down

0 comments on commit b260d50

Please sign in to comment.