Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llamamodel: add 12 new architectures for CPU (and Metal) inference #1914

Merged
merged 1 commit into from
Feb 5, 2024

Conversation

cebtenzzre
Copy link
Member

@cebtenzzre cebtenzzre commented Feb 2, 2024

This PR enables support for the following model architectures:

  • Baichuan
  • BLOOM
  • CodeShell
  • GPT-2
  • Orion
  • Persimmon
  • Phi and Phi-2
  • Plamo
  • Qwen
  • Qwen2
  • Refact
  • StableLM

These are only in the CPU/Metal inference whitelist until we know that any of them are supported on Vulkan.

- Baichuan
- BLOOM
- CodeShell
- GPT-2
- Orion
- Persimmon
- Phi and Phi-2
- Plamo
- Qwen
- Qwen2
- Refact
- StableLM

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
@cebtenzzre cebtenzzre merged commit 92c025a into main Feb 5, 2024
6 of 10 checks passed
@cebtenzzre cebtenzzre deleted the add-new-arches branch February 5, 2024 21:49
@cebtenzzre cebtenzzre changed the title llamamodel: add 12 new architectures for CPU inference llamamodel: add 12 new architectures for CPU (and Metal) inference Feb 8, 2024
dpsalvatierra pushed a commit to dpsalvatierra/gpt4all that referenced this pull request Feb 16, 2024
Baichuan, BLOOM, CodeShell, GPT-2, Orion, Persimmon, Phi and Phi-2,
Plamo, Qwen, Qwen2, Refact, StableLM

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants