Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enable bf16 output in TBE CPU kernel for other input types #1851

Closed
wants to merge 1 commit into from

Conversation

excelle08
Copy link

Summary: Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

@netlify
Copy link

netlify bot commented Jun 26, 2023

Deploy Preview for pytorch-fbgemm-docs canceled.

Name Link
🔨 Latest commit 374b556
🔍 Latest deploy log https://app.netlify.com/sites/pytorch-fbgemm-docs/deploys/64a7692515b2a200089456e6

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jun 26, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: b49b8f8bfc6932c5665daa4b810fcea1f306a077
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jun 27, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: b52a672c7aeb175746b62ea6dfbbad81eec14975
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jun 28, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: 73802deb5e17d4da4b84c7d75c6a5c3c415b46d6
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jun 28, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: 94c48bd4fc68da13505d8080c5773e214da0fef4
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jun 30, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: 6f759ef4b072e6e8b9d78b550497c1ebe1b3483d
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jun 30, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: 8721ba1aa097702ae6a0844d312929124ed3448e
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 5, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: cc09c0c60cca6f4b72036d2e28e452f4f9334b5c
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 5, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: 14f0a8a33b906ac531a8bb5d212f74b7eb2cf00d
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 5, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: aca192f78900a09eee6cda2d660cf4761c765deb
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 5, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: 61301ea038ffa358550dccc346e4df3096781901
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 6, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Differential Revision: D47028021

fbshipit-source-id: d8b83bb3fb4c477c361a1f062e3c254d840e01d2
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 7, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Reviewed By: sryap

Differential Revision: D47028021

fbshipit-source-id: 26bd071461ae65ceafc0d27c2233ec340bb24180
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

excelle08 pushed a commit to excelle08/FBGEMM that referenced this pull request Jul 7, 2023
)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Reviewed By: sryap

Differential Revision: D47028021

fbshipit-source-id: b71c123e4f76d92f0e85bb0e00512b57a02695c5
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

)

Summary:
Pull Request resolved: pytorch#1851

Enable bf16 output support in TBE CPU kernel when the input weight type is int8/fp8/fp16/fp32

Reviewed By: sryap

Differential Revision: D47028021

fbshipit-source-id: 8c3f09113541ce813a424cdf2f39a4679cd2115d
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47028021

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in ed4fe6e.

pytorchmergebot pushed a commit to pytorch/pytorch that referenced this pull request Oct 21, 2023
By updating fbgemm submodule
Add regression test for it (though it can probably be limited to just CPU as reproducer only works if num_threads is 1)

Also, update call-sites  to `fbgemm:: GenerateEmbeddingSpMDM` to pass `isbf16` twice, to match API changes introduced in pytorch/FBGEMM#1851

Fixes #111189 and #111710

Pull Request resolved: #111672
Approved by: https://github.com/Skylion007
andreigh pushed a commit to andreigh/pytorch that referenced this pull request Oct 26, 2023
By updating fbgemm submodule
Add regression test for it (though it can probably be limited to just CPU as reproducer only works if num_threads is 1)

Also, update call-sites  to `fbgemm:: GenerateEmbeddingSpMDM` to pass `isbf16` twice, to match API changes introduced in pytorch/FBGEMM#1851

Fixes pytorch#111189 and pytorch#111710

Pull Request resolved: pytorch#111672
Approved by: https://github.com/Skylion007
xuhancn pushed a commit to xuhancn/pytorch that referenced this pull request Nov 7, 2023
By updating fbgemm submodule
Add regression test for it (though it can probably be limited to just CPU as reproducer only works if num_threads is 1)

Also, update call-sites  to `fbgemm:: GenerateEmbeddingSpMDM` to pass `isbf16` twice, to match API changes introduced in pytorch/FBGEMM#1851

Fixes pytorch#111189 and pytorch#111710

Pull Request resolved: pytorch#111672
Approved by: https://github.com/Skylion007
Skylion007 pushed a commit to Skylion007/pytorch that referenced this pull request Nov 14, 2023
By updating fbgemm submodule
Add regression test for it (though it can probably be limited to just CPU as reproducer only works if num_threads is 1)

Also, update call-sites  to `fbgemm:: GenerateEmbeddingSpMDM` to pass `isbf16` twice, to match API changes introduced in pytorch/FBGEMM#1851

Fixes pytorch#111189 and pytorch#111710

Pull Request resolved: pytorch#111672
Approved by: https://github.com/Skylion007
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants