Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

update quantization doc: add x86 backend as default backend of server inference #86794

Closed

Conversation

XiaobingSuper
Copy link
Collaborator

@XiaobingSuper XiaobingSuper commented Oct 12, 2022

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 12, 2022

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/86794

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit c3d8613:
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

XiaobingSuper added a commit that referenced this pull request Oct 12, 2022
ghstack-source-id: e3f8c232ce566213b31fd5448a07a86ca41a0d3f
Pull Request resolved: #86794
XiaobingSuper added a commit that referenced this pull request Oct 12, 2022
ghstack-source-id: 4bb0ca12ffe3465d9970b4c6698fac29c9bc6289
Pull Request resolved: #86794
@jgong5
Copy link
Collaborator

jgong5 commented Oct 13, 2022

Since x86 is now the default qengine replacing fbgemm, can we only recommend x86 for server CPUs? We can leave a note that fbgemm is still available but not recommended.

XiaobingSuper added a commit that referenced this pull request Oct 13, 2022
ghstack-source-id: 08dd01ebb39df926006c3126cde54d300844381b
Pull Request resolved: #86794
XiaobingSuper added a commit that referenced this pull request Oct 13, 2022
ghstack-source-id: 5a2fbb19f228039521387b824bd9c35bba7aca62
Pull Request resolved: #86794
@@ -742,30 +748,31 @@ Backend/Hardware Support

Today, PyTorch supports the following backends for running quantized operators efficiently:

* x86 CPUs with AVX2 support or higher (without AVX2 some operations have inefficient implementations), via `fbgemm <https://github.com/pytorch/FBGEMM>`_
* x86 CPUs with AVX2 support or higher (without AVX2 some operations have inefficient implementations), via `x86` to apply the optimization of `fbgemm <https://github.com/pytorch/FBGEMM>`_ and `onednn <https://github.com/oneapi-src/oneDNN>`_ (see the details at `RFC <https://github.com/pytorch/pytorch/issues/83888>`_)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
* x86 CPUs with AVX2 support or higher (without AVX2 some operations have inefficient implementations), via `x86` to apply the optimization of `fbgemm <https://github.com/pytorch/FBGEMM>`_ and `onednn <https://github.com/oneapi-src/oneDNN>`_ (see the details at `RFC <https://github.com/pytorch/pytorch/issues/83888>`_)
* x86 CPUs with AVX2 support or higher (without AVX2 some operations have inefficient implementations), via `x86` optimized by `fbgemm <https://github.com/pytorch/FBGEMM>`_ and `onednn <https://github.com/oneapi-src/oneDNN>`_ (see the details at `RFC <https://github.com/pytorch/pytorch/issues/83888>`_)

XiaobingSuper added a commit that referenced this pull request Oct 13, 2022
ghstack-source-id: 4f27ab24f6341b4f69d3a98a396ba4e070f2fa2f
Pull Request resolved: #86794
@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Oct 13, 2022
@XiaobingSuper XiaobingSuper changed the title update quantization doc: add onednn and x86 backend description update quantization doc: add x86 backend as default backend of server inference Oct 13, 2022
@kit1980
Copy link
Member

kit1980 commented Nov 18, 2022

@pytorchbot rebase

@kit1980
Copy link
Member

kit1980 commented Nov 18, 2022

@XiaobingSuper feel free to merge when ready.

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a rebase job. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Rebase failed due to Command git -C /home/runner/work/pytorch/pytorch rebase refs/remotes/origin/viable/strict gh/XiaobingSuper/16/orig returned non-zero exit code 1

Rebasing (1/1)
Auto-merging docs/source/quantization.rst
CONFLICT (content): Merge conflict in docs/source/quantization.rst
error: could not apply 4ae3e0ee3c... update quantization doc: add onednn and x86 backend description
hint: Resolve all conflicts manually, mark them as resolved with
hint: "git add/rm <conflicted_files>", then run "git rebase --continue".
hint: You can instead skip this commit: run "git rebase --skip".
hint: To abort and get back to the state before "git rebase", run "git rebase --abort".
Could not apply 4ae3e0ee3c... update quantization doc: add onednn and x86 backend description

Raised by https://github.com/pytorch/pytorch/actions/runs/3493543061

XiaobingSuper added a commit that referenced this pull request Dec 1, 2022
ghstack-source-id: 6e9d7bef60d4f6312c66971b9643bcbcedfc05df
Pull Request resolved: #86794
@XiaobingSuper XiaobingSuper added the release notes: quantization release notes category label Dec 1, 2022
@XiaobingSuper
Copy link
Collaborator Author

@kit1980 @jerryzh168, code is rebased.

@XiaobingSuper
Copy link
Collaborator Author

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/trunk Trigger trunk jobs on your pull request Merged open source release notes: quantization release notes category
Projects
Status: Done
Development

Successfully merging this pull request may close these issues.

None yet

5 participants