Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for QWEN and Baichuan2 models #1731

Closed
sorasoras opened this issue Dec 8, 2023 · 5 comments
Closed

Support for QWEN and Baichuan2 models #1731

sorasoras opened this issue Dec 8, 2023 · 5 comments
Labels
backend gpt4all-backend issues models

Comments

@sorasoras
Copy link

Feature request

recently, https://github.com/ggerganov/llama.cpp has add support for both QWEN and Baichuan2.
It has added QWEN at 1610.
ggerganov/llama.cpp#4281
I have look up the Nomic Vulkan Fork of LLaMa.cpp,
it does have support for Baichuan2 but not QWEN, but GPT4ALL itself does not support Baichuan2.

Motivation

I failed to load baichuan2 and QWEN models, GPT4ALL supposed to be easy to use

Your contribution

Not quite as i am not a programmer but i would look up if that helps

@cebtenzzre cebtenzzre added backend gpt4all-backend issues models labels Dec 10, 2023
@cebtenzzre
Copy link
Member

Baichuan2 should be supported in the next release - the current release isn't using the latest version of our llama.cpp fork.

Qwen got merged upstream a little too late, but it should be supported here soon.

@AdkinsHan
Copy link

There is a problem. When I use Qwen, no matter the graphics card or CPU specified by my device, it has been calling the CPU and memory for calculation instead of calling the GPU.
mode_name="qwen1_5-14b-chat-q8_0.gguf"
gpt4all version 2.2.1.post1
https://github.com/QwenLM/Qwen
image
image
When I use it, I see that the memory and CPU are being called, but the video memory usage has not changed.

@AdkinsHan
Copy link

When I use other GGUF models I can see the following output
llama.cpp: using Vulkan on NVIDIA GeForce RTX 3070
image

@cebtenzzre
Copy link
Member

cebtenzzre commented Feb 22, 2024

As of #2005, Qwen, Qwen2, and Baichuan all have GPU support enabled on both the Metal and Vulkan backends.

@cebtenzzre cebtenzzre added the awaiting-release issue is awaiting next release label Feb 22, 2024
@cebtenzzre
Copy link
Member

This is implemented in the v2.7.2 release.

@cebtenzzre cebtenzzre removed the awaiting-release issue is awaiting next release label Mar 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
backend gpt4all-backend issues models
Projects
None yet
Development

No branches or pull requests

3 participants