-
Notifications
You must be signed in to change notification settings - Fork 174
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tensor Core Layout docs is not clear #386
Comments
|
@jerryzh168 is this being addressed by @jainapurva refactor or still an issue? |
not fully I think, but I can follow up with some doc fixes to address this |
jerryzh168
added a commit
to jerryzh168/ao
that referenced
this issue
Oct 18, 2024
Summary: Following pytorch#988 we added TP support for int4_weight_only quantization in torchao that's using TensorCoreTiledLayout Addresses one work item in pytorch#988 Also clarified docs based on pytorch#386 Also restructructured the tests in test/dtypes/test_affine_quantized_tensor_parallel.py to not depend on torchao/utils.py to reduce the jumps people have to do to understand what is tested Test Plan: python test/dtypes/test_affine_quantized_tensor_parallel.py Reviewers: Subscribers: Tasks: Tags:
jerryzh168
added a commit
that referenced
this issue
Oct 19, 2024
* Add tensor parallelism support for int4_weight_only quantization Summary: Following #988 we added TP support for int4_weight_only quantization in torchao that's using TensorCoreTiledLayout Addresses one work item in #988 Also clarified docs based on #386 Also restructructured the tests in test/dtypes/test_affine_quantized_tensor_parallel.py to not depend on torchao/utils.py to reduce the jumps people have to do to understand what is tested Test Plan: python test/dtypes/test_affine_quantized_tensor_parallel.py Reviewers: Subscribers: Tasks: Tags: * typo
I think we can close now, I just landed #1120, cc @msaroufim please let me know if there is any follow up questions around tensor core tiled layout or layout |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Right now what we have is docstrings but they could use work - this came up as @vayuda was looking at extending his bitpacking work to include a notion of scales
torch.ops.aten._weight_int4pack_mm(input_tensor.contiguous(), packed_weight, groupsize, scale_and_zero)
unclear why scale_and_zero
are a single tensorinnerKtiles
is never definedThe text was updated successfully, but these errors were encountered: