-
Notifications
You must be signed in to change notification settings - Fork 762
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.
Already on GitHub? Sign in to your account
Stability AI's Stable Code 3B support #1230
Comments
Since stability ai folks provides gguf quantization, it's easy to integrate by following: https://slack.tabbyml.com/Gd5zV1P69JN/how-can-i-indicate-a-custom-model-to-tabbyml |
For running tabby on downloaded model, you could refer https://github.com/TabbyML/tabby/blob/main/MODEL_SPEC.md directly |
Can the gguf file be located elsewhere other than in the Tabby model file folder? Want to use the gguf file LM Studio has already downloaded to save disk space. |
Yes, but you still need to organize it into a directory format specified in model spec.
Shall be good to create a symbolic link? |
So if I:
it should work? |
Apparently not. Deletes model rows to all local file. thread 'main' panicked at crates/tabby-common/src/registry.rs:87:9: |
Hi, could you share:
|
URL to original file URL to symbolic link RUST_BACKTRACE=full tabby serve --device metal --model TabbyML/DeepseekCoder-6.7B/ |
|
I'm not able to load the model. I tried the q8 quantised files of both https://huggingface.co/brittlewis12/stable-code-3b-GGUF and https://huggingface.co/TheBloke/stable-code-3b-GGUF but llama.cpp is unable to load the file.
I also tried to convert the GGUF model again using llama.cpp, in case the format has changed, but this wasn't helpful.
|
thread 'main' panicked at crates/tabby-common/src/registry.rs:87:9: |
I get the same as @fungiboletus with my custom registry. tabby serve --device metal --model anoldguy/StableCode-3B
Writing to new file.
馃幆 Downloaded https://huggingface.co/stabilityai/stable-code-3b/resolve/main/stable-code-3b-Q6_K.gguf to /Users/nathan/.tabby/models/anoldguy/StableCode-3B/ggml/q8_0.v2.gguf.tmp
00:00:23 鈻曗枅鈻堚枅鈻堚枅鈻堚枅鈻堚枅鈻堚枅鈻堚枅鈻堚枅鈻堚枅鈻堚枅鈻堚枏 2.14 GiB/2.14 GiB 93.90 MiB/s ETA 0s. 2024-01-18T14:05:47.674093Z INFO tabby::serve: crates/tabby/src/serve.rs:111: Starting server, this might takes a few minutes...
2024-01-18T14:05:47.674908Z INFO tabby::services::code: crates/tabby/src/services/code.rs:53: Index is ready, enabling server...
2024-01-18T14:05:47.772956Z ERROR llama_cpp_bindings: crates/llama-cpp-bindings/src/lib.rs:62: Unable to load model: /Users/nathan/.tabby/models/anoldguy/StableCode-3B/ggml/q8_0.v2.gguf I'm using this as the definition, but I'm unsure about the prompt template. 馃 {
"name": "StableCode-3B",
"license_name": "STABILITY AI NON-COMMERCIAL RESEARCH COMMUNITY LICENSE",
"license_url": "https://huggingface.co/stabilityai/stable-code-3b/blob/main/LICENSE",
"prompt_template": "<fim_prefix>{prefix}<fim_suffix>{suffix}<fim_middle>",
"provider_url": "https://huggingface.co/stabilityai/stable-code-3b",
"urls": [
"https://huggingface.co/stabilityai/stable-code-3b/resolve/main/stable-code-3b-Q6_K.gguf"
],
"sha256": "9749daf176491c33a7318660f1637c97674b0070d81740be8763b2811c495bfc"
} |
Thanks for doing the experiment, it seems the reason that stable lm support is added after our current checkpoint of llama.cpp. this should be fixed after #434 is done |
Is there any plan to support Stable Code 3B officially in the future? |
We've bumped llama.cpp version and it has been released in https://github.com/TabbyML/tabby/releases/tag/nightly Please give it a try to see if it works with StableCode-3B |
Fixed in v0.8.0 |
Please describe the feature you want
Please add Stability AI's Stable Code 3B
https://huggingface.co/stabilityai/stable-code-3b
Please reply with a 馃憤 if you want this feature.
The text was updated successfully, but these errors were encountered: