Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama_state #62

Closed
martindevans opened this issue Jul 24, 2023 · 2 comments
Closed

llama_state #62

martindevans opened this issue Jul 24, 2023 · 2 comments

Comments

@martindevans
Copy link
Member

It looks like recent updates to Llama.cpp (e.g. ggerganov/llama.cpp#1797) have modified the API significantly with regards to how "state" is handled.

The llama_model is loaded with one API call (llama_load_model_from_file), which loads all of the static data (weights, vocabulary etc) and then you can create one or more states over this (llama_new_context_with_model).

Is anyone else working on this? If not I'm happy to have a go at it.

@martindevans
Copy link
Member Author

Low level changes addressed in #64

@martindevans
Copy link
Member Author

Some higher level changes addressed in #70

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant