Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FEAT: support fp4 and int8 quantization for pytorch model #238

Merged
merged 4 commits into from
Jul 26, 2023

Conversation

pangyoki
Copy link
Contributor

@pangyoki pangyoki commented Jul 24, 2023

issue #230 and #239

@XprobeBot XprobeBot added this to the v0.1.0 milestone Jul 24, 2023
@pangyoki pangyoki force-pushed the pytorch_support_quantization branch from f752c49 to 6c79197 Compare July 25, 2023 04:11
@pangyoki pangyoki force-pushed the pytorch_support_quantization branch from 37b69d1 to 5be5e5a Compare July 25, 2023 06:30
xinference/model/llm/pytorch/core.py Outdated Show resolved Hide resolved
xinference/model/llm/pytorch/core.py Outdated Show resolved Hide resolved
@UranusSeven UranusSeven merged commit e4115d1 into xorbitsai:main Jul 26, 2023
9 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants