Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Custom GGML outside LlamaCpp scope #38

Open
su77ungr opened this issue May 14, 2023 · 6 comments
Open

Custom GGML outside LlamaCpp scope #38

su77ungr opened this issue May 14, 2023 · 6 comments
Labels
enhancement New feature or request model

Comments

@su77ungr
Copy link
Owner

For the MosaiML: haven't tried yet, feel free to create another issue so that we don't forget after closing this one
Update: mpt-7b-q4_0.bin doesn't work "out of the box", it yields what(): unexpectedly reached end of file and a runtime error.

Originally posted by @hippalectryon-0 in #33 (comment)

@su77ungr su77ungr added the enhancement New feature or request label May 14, 2023
@su77ungr
Copy link
Owner Author

Outsourced curated list of supported models; later adding to README.md

@su77ungr su77ungr added the model label May 14, 2023
@su77ungr
Copy link
Owner Author

su77ungr commented May 14, 2023

Maye create setup.py that fetches directly from HF

Edit: this does counteract the air-gapped idea

from huggingface_hub import hf_hub_download

#Download the model
hf_hub_download(repo_id="LLukas22/gpt4all-lora-quantized-ggjt", filename="ggjt-model.bin", local_dir=".")

Edit: implemented with #61
Also @hippalectryon-0 did you test the 4bit or 16 from Mosaic?

@su77ungr su77ungr pinned this issue May 14, 2023
@hippalectryon-0
Copy link
Contributor

Only mpt-7b-q4_0.bin from https://huggingface.co/LLukas22/mpt-7b-ggml

@imSrbh
Copy link

imSrbh commented May 15, 2023

I feel this mpt-7B is faster than the existing one here.

@su77ungr
Copy link
Owner Author

su77ungr commented May 15, 2023

You got it running? We should add benchmark runs so everyone can plot and share results.

@hippalectryon-0
Copy link
Contributor

ggerganov/llama.cpp#1333

@su77ungr su77ungr unpinned this issue May 18, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request model
Projects
None yet
Development

No branches or pull requests

3 participants