Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Request for the Llama-2-13B with AQLM (2x8 scheme) #90

Closed
ChenMnZ opened this issue May 18, 2024 · 3 comments
Closed

Request for the Llama-2-13B with AQLM (2x8 scheme) #90

ChenMnZ opened this issue May 18, 2024 · 3 comments

Comments

@ChenMnZ
Copy link

ChenMnZ commented May 18, 2024

Hello,

Thanks for your outstanding works. I want to do a compressive comparison of recent quantization methods.

Due to that latest lm-eval can obtain higher accuracy than the ones reported in paper, I have to re-evaluate each quantized models.

I found that there is not model about Llama-2-13B with AQLM (2x8 scheme), can you share it on huggingface?

Thank you!

@Vahe1994
Copy link
Owner

Hello!
Sorry for late response. It was very busy several weeks.
Thanks for interest in the work. Currently 13B 2x8gs8 quantization is running. When it is done, I will put it in HF hub and tell you.

P.s. Additionally you can check out this PR #93 for new LM eval code. Thanks to fast dequantization kernels, you can run lm eval from HF as any other models.
For ppl please check out #91.

@Vahe1994
Copy link
Owner

Vahe1994 commented Jun 6, 2024

@Vahe1994 Vahe1994 closed this as completed Jun 6, 2024
@Vahe1994 Vahe1994 reopened this Jun 6, 2024
@ChenMnZ
Copy link
Author

ChenMnZ commented Jun 6, 2024

Thanks for your time!!! I will try this model.

@ChenMnZ ChenMnZ closed this as completed Jun 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants