Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

gguf : add BERT, MPT, and GPT-J arch info #3408

Merged
merged 2 commits into from
Oct 2, 2023

Conversation

cebtenzzre
Copy link
Collaborator

These GGUF architectures will be used in a future release of gpt4all.

@mchiang0610
Copy link

mchiang0610 commented Sep 30, 2023

This is awesome! Will you be adding these to the implementation of these architectures or will it just be used for your private fork?

Support for these architectures would be amazing for the community as a whole.

@cebtenzzre
Copy link
Collaborator Author

The conversion scripts and CPU inference implementations are here: https://github.com/nomic-ai/gpt4all/tree/gguf_latest_llama/gpt4all-backend

I was mainly focused on updating the existing gpt4all code, so there are surely improvements from ggml and koboldcpp that have not been included.

Which model architecture would be best to add upstream support for first? There is also GPT-NeoX, which is not part of gpt4all.

@cebtenzzre cebtenzzre changed the title gguf : add BERT, MPT, and GPT-J model architectures gguf : add BERT, MPT, and GPT-J arch info Sep 30, 2023
@cebtenzzre cebtenzzre merged commit 29a404a into ggerganov:master Oct 2, 2023
9 of 10 checks passed
joelkuiper added a commit to vortext/llama.cpp that referenced this pull request Oct 5, 2023
…example

* 'master' of github.com:ggerganov/llama.cpp: (24 commits)
  convert : fix Baichuan2 models by using vocab size in config.json (ggerganov#3299)
  readme : add project status link
  ggml : fix build after ggerganov#3329
  llm : add Refact model (ggerganov#3329)
  sync : ggml (conv 1d + 2d updates, UB fixes) (ggerganov#3468)
  finetune : readme fix typo (ggerganov#3465)
  ggml : add RISC-V Vector Support for K-Quants and improved the existing intrinsics (ggerganov#3453)
  main : consistent prefix/suffix coloring (ggerganov#3425)
  llama : fix session saving/loading (ggerganov#3400)
  llama : expose model's rope_freq_scale in the API (ggerganov#3418)
  metal : alibi for arbitrary number of heads (ggerganov#3426)
  cmake : make LLAMA_NATIVE flag actually use the instructions supported by the processor (ggerganov#3273)
  Work on the BPE tokenizer (ggerganov#3252)
  convert : fix vocab size when not defined in hparams (ggerganov#3421)
  cmake : increase minimum version for add_link_options (ggerganov#3444)
  CLBlast: Add broadcast support for matrix multiplication (ggerganov#3402)
  gguf : add BERT, MPT, and GPT-J arch info (ggerganov#3408)
  gguf : general usability improvements (ggerganov#3409)
  cmake : make CUDA flags more similar to the Makefile (ggerganov#3420)
  finetune : fix ggerganov#3404 (ggerganov#3437)
  ...
yusiwen pushed a commit to yusiwen/llama.cpp that referenced this pull request Oct 7, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants