-
Notifications
You must be signed in to change notification settings - Fork 817
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama.cpp not working and got errors (doc is correct???) #445
Comments
Thanks for sharing these, it's nothing related to the config, these were bugs, and I've been able to fix them already. I will upload a new version very soon and let you know when it's ready |
@kun432 The newest version is ready, if you update the extension to v0.0.359 this problem should no longer occur |
@sestinj Thank you! I updated the latest version and errors gone. but no output... but I added models=Models(
default=LlamaCpp(
max_context_length=16384,
server_url="http://192.168.1.1:8080",
llama_cpp_args={'stop': ['[INST]']}
)
), is this the expecetd behavior? If so, doc should be updated. |
Looks like this was because of a mistake I made in the default llama_cpp_args, and so setting them overrode the mistake. So no, you should not have to do this. I'll make an update with the fix. Thanks for catching this! |
New version is ready, should be able to leave out llama_cpp_args now |
Great, works in v0.0.364! will explore other features in Continue! Thanks! |
awesome! let me know what other questions you run into. Feel free to use Discord if it's ever more convenient: https://discord.gg/NWtdYexhMs |
Describe the bug
I want to use Code Llama via llama.cpp's HTTP server. Setting up accoringly to the doc, but got 2 errors:
To Reproduce
Steps to reproduce the behavior:
config.py
. Only the following points changed from the default:Environment
Logs
See above.
Additional context
I also read #410 and tried some changes but no lucks. I just want to know the correct way to set config for Llama.cpp.
The text was updated successfully, but these errors were encountered: