New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is GGUF support coming anytime soon ? #1435
Comments
Yes, it is the next highest priority goal after getting a nightly build and I'll be working on it this coming week. |
Keep up your good word, manyoso! |
How progress with GGUF support is going? |
GGUF support is currently available in this pre-release, which is available for beta testing: |
GGUF "only" ? |
It's a lot of effort to maintain support for older file formats, since the current version of llama.cpp only supports GGUF. GGUF is intended to be a long-term solution that can easily be extended, so it should be the last new format for a long time. There is a script in llama.cpp that you can use to convert LLaMA models to the new format, called 'convert-llama-ggml-to-gguf.py'. Many of the models previously available in GPT4All are available for download in the new format, and many GGML models quantized by TheBloke have GGUF variants now as well. |
* Added models for gpt4all. * Removed replit model. * Updated docs for gpt4all. * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Updated docs. * A fix to stop download of embeddings model. --------- Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Feature request
Is GGUF support coming anytime soon ? Nobody seems to be doing GGML anymore as the project was phased out about a month ago.
Motivation
Yes
Your contribution
What?
The text was updated successfully, but these errors were encountered: