-
Notifications
You must be signed in to change notification settings - Fork 773
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support starcoder2 #1587
Comments
ref: #1230 |
also pending on ggerganov/llama.cpp#5795 |
added https://github.com/wsxiaoys/registry-tabby, runnable with |
@wsxiaoys having trouble getting it to run , i can run the normal models on the nightly build , but for some reason can't get starcoder2 to work, used your repo and my fork , also have tried manually downloading it. https://huggingface.co/brittlewis12/starcoder2-3b-GGUF |
Hey @aaronstevenson408 , could you try this docker images |
Seems like it worked ty |
Is the prompt_template the same has starcoder 1? |
I am not sure if here is the right place to report this, but the three size for StarCoder2 all have the name "StarCoder2-3B" there. |
Thanks for reporting - fixed |
exit ....
|
GGUF.... |
Hi @wsxiaoys, is it now possible to run StarCoder v2 from tabby directly with something like: |
@jeromepl you can you can either make you own repo registry-tabby or use one already available
|
Is this still pending to be implemented on the official model registry? |
Just added the Starcoder2-3B / Starcoder2-7B to the official registry. Enjoy! |
@wsxiaoys could you add the 15B? It's actually quite usable on a 4090 or a 7900XTX. |
@rudiservo you could still maintain that in a forked registry. For tabby official registry we prefer not to include models go beyond 10B atm. |
@wsxiaoys yes I already have one, I was just suggesting it since more people might be interested in it and given my experience with the model, also the quality between 7B and 15B is a bit noticeable. It is quite usable with Tabby, but you do need more than 16GB of VRAM, currently the 7900XTX can handle it quite well. |
I see StarCoder2-7B and StarCoder2-15B in the Tabby leaderboard, but only 7B is available on Mac: tabby serve --device metal --model StarCoder2-15B |
It looks like the StarCoder2 15B quantized model is available in HuggingFace: |
https://techcrunch.com/2024/02/28/starcoder-2-is-a-code-generating-ai-that-runs-on-most-gpus
https://github.com/bigcode-project/starcoder2
The text was updated successfully, but these errors were encountered: