Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 0 additions & 3 deletions backends/xnnpack/utils/quant_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -222,9 +222,6 @@ def extract_qdq_affine_op_args_for_decomposed_ops(node: torch.fx.Node):

# add target_dtype_node after quant_min/quant_max
args.append(target_dtype)
# zero_point_domain
if len(node.args) > 7 and node.args[7] != "INT":
return None, None

if is_per_channel_group(node):
block_sizes = cast(list[int], node.args[1])
Expand Down
3 changes: 0 additions & 3 deletions exir/passes/_quant_patterns_and_replacements.py
Original file line number Diff line number Diff line change
Expand Up @@ -1017,7 +1017,6 @@ def embedding_byte_dtype_pattern(
torch.int8,
-128,
127,
"INT",
output_dtype,
)
return torch.ops.aten.embedding.default(dq, indices)
Expand Down Expand Up @@ -1062,7 +1061,6 @@ def embedding_2bit_dtype_pattern(
torch.int8,
-2,
1,
"INT",
output_dtype,
)
return torch.ops.aten.embedding.default(dq, indices)
Expand Down Expand Up @@ -1110,7 +1108,6 @@ def embedding_4bit_dtype_pattern(
torch.int8,
-8,
7,
"INT",
output_dtype,
)
return torch.ops.aten.embedding.default(dq, indices)
Expand Down
Loading