Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to run LLama-3 or Phi with more then 4096 prompt tokens? #2171

Open
baleksey opened this issue May 7, 2024 · 0 comments
Open

How to run LLama-3 or Phi with more then 4096 prompt tokens? #2171

baleksey opened this issue May 7, 2024 · 0 comments

Comments

@baleksey
Copy link

baleksey commented May 7, 2024

Could you please show me an example where LLama-3 model used (better GGUF quantized) and initial prompt is more then 4096 tokens long? Or better 16-64K long (for RAG). Currently everything I do ends with error:
In this code:
let logits = model.forward(&input, 0); // input is > 4096 tokens

Error:
narrow invalid args start + len > dim_len: [4096, 64], dim: 0, start: 0, len:4240

Model used:
https://huggingface.co/MaziyarPanahi/Llama-3-8B-Instruct-64k-GGUF

Thank you a lot in advance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant