Skip to content

Commit

Permalink
#8364: Revert "skip concat tests which fallback to torch"
Browse files Browse the repository at this point in the history
This reverts commit bf163e2.
  • Loading branch information
ayerofieiev-tt committed May 15, 2024
1 parent f4e1c13 commit 47634f3
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 8 deletions.
5 changes: 0 additions & 5 deletions tests/ttnn/unit_tests/operations/test_concat.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,9 +15,6 @@
@pytest.mark.parametrize("width", [4, 32])
@pytest.mark.parametrize("dim", [0, 1])
def test_concat(device, height, width, dim):
if height % ttnn.TILE_SIZE != 0 or width % ttnn.TILE_SIZE != 0:
pytest.skip("ttnn.concat only supports tensors with Layout.TILE_LAYOUT without a padding")

torch_input_tensor_a = torch.rand((height, width), dtype=torch.bfloat16)
torch_input_tensor_b = torch.rand((height, width), dtype=torch.bfloat16)
torch_output_tensor = torch.concat([torch_input_tensor_a, torch_input_tensor_b], dim=dim)
Expand Down Expand Up @@ -81,8 +78,6 @@ def test_concat(device, height, width, dim):
def test_sharded_concat(
device, input_shape_a, shard_shape_a, input_shape_b, shard_shape_b, output_shard_shape, shard_grid
):
pytest.skip("ttnn.concat only supports tensors with Layout.TILE_LAYOUT without a padding")

input_a_sharded_memory_config = ttnn.create_sharded_memory_config(
shard_shape_a,
core_grid=shard_grid,
Expand Down
4 changes: 1 addition & 3 deletions ttnn/ttnn/operations/data_movement.py
Original file line number Diff line number Diff line change
Expand Up @@ -328,9 +328,7 @@ def concat(
output_tensor = ttnn.squeeze(output_tensor, dim=0)
return output_tensor
else:
raise NotImplementedError(
"ttnn.concat only supports tensors with Layout.TILE_LAYOUT without a padding, with rank <= 4"
)
raise NotImplementedError


def _golden_function(input_tensor, split_size, dim):
Expand Down

0 comments on commit 47634f3

Please sign in to comment.