New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CPU instead of GPU #4
Comments
You won't be able to do that with this particular codebase (which is heavily dependent on FasterTransformer so that it can get decent latency), but this gist from @nforest shows how to use any model from Huggingface Transformers with Copilot, and Transformers supports doing CPU-only: https://gist.github.com/nforest/d1432b917468f5ad24b83954c98e67b1 You should be able to pass in I do warn you that it will be a lot slower! |
Thank you! Edit: Running off CPU makes this more accessible by moving the barrier to entry from a $3000 rig to a $1000 rig. |
@TechnologyClassroom Did you tried it? How long did it take for a completion on what cpu? 😃 |
@1muen I didn't try it yet. I commented asking about licensing and have not heard back. (That project really deserves to be a full repo too) |
Is there a way to do such a thing with CPU instead of GPU? I know this would be slower, but it would be a cheaper solution and would not depend on NVIDIA.
The text was updated successfully, but these errors were encountered: