Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is GGUF support coming anytime soon ? #1435

Closed
PedzacyKapec opened this issue Sep 17, 2023 · 6 comments
Closed

Is GGUF support coming anytime soon ? #1435

PedzacyKapec opened this issue Sep 17, 2023 · 6 comments
Labels
backend gpt4all-backend issues

Comments

@PedzacyKapec
Copy link

Feature request

Is GGUF support coming anytime soon ? Nobody seems to be doing GGML anymore as the project was phased out about a month ago.

Motivation

Yes

Your contribution

What?

@manyoso
Copy link
Collaborator

manyoso commented Sep 17, 2023

Yes, it is the next highest priority goal after getting a nightly build and I'll be working on it this coming week.

@TheCopacabanaMan
Copy link

Keep up your good word, manyoso!
You, and all your team!

@JamshedQurbonboev
Copy link

How progress with GGUF support is going?

@cebtenzzre
Copy link
Member

GGUF support is currently available in this pre-release, which is available for beta testing:
https://github.com/nomic-ai/gpt4all/releases/tag/v2.5.0-pre1

@PedzacyKapec
Copy link
Author

GGUF support is currently available in this pre-release, which is available for beta testing: https://github.com/nomic-ai/gpt4all/releases/tag/v2.5.0-pre1

GGUF "only" ?
What do you mean ? No GGML support ?
WHy cant both be supported ?

@cebtenzzre
Copy link
Member

WHy cant both be supported ?

It's a lot of effort to maintain support for older file formats, since the current version of llama.cpp only supports GGUF. GGUF is intended to be a long-term solution that can easily be extended, so it should be the last new format for a long time.

There is a script in llama.cpp that you can use to convert LLaMA models to the new format, called 'convert-llama-ggml-to-gguf.py'. Many of the models previously available in GPT4All are available for download in the new format, and many GGML models quantized by TheBloke have GGUF variants now as well.

@cebtenzzre cebtenzzre added backend gpt4all-backend issues awaiting-release issue is awaiting next release labels Oct 11, 2023
@cebtenzzre cebtenzzre removed the awaiting-release issue is awaiting next release label Oct 24, 2023
JasonWeill referenced this issue in jupyterlab/jupyter-ai Nov 10, 2023
* Added models for gpt4all.

* Removed replit model.

* Updated docs for gpt4all.

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Updated docs.

* A fix to stop download of embeddings model.

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
backend gpt4all-backend issues
Projects
None yet
Development

No branches or pull requests

5 participants