-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use llama.cpp or TextGen-webui? #1
Comments
Great lmk how that goes I’d love to make it so you can use any model you
want. I used the llama 2 chat just to show how good this can work with a
limited non fine tuned instruction model and it’s actually not bad. Any
ideas for some 13b or less models that could write react really well? I’ll
just change that as the default.
On Mon, Aug 21, 2023 at 4:54 PM Michael Martinez ***@***.***> wrote:
Cool project!!!
There are a lot of models out there that will probably perform way better
than vanilla llama 2. To get an idea, have a look at this HF space:
https://huggingface.co/spaces/mike-ravkine/can-ai-code-results
Also, using a locally run 13b model is relatively trivial at this point,
even with modest hardware if you use a quantized version. I am looking your
code base to see where a good entry point for something like llama.cpp or
textgen-webui could be harnessed.
—
Reply to this email directly, view it on GitHub
<#1>, or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAPIGZ5JKFINUNBVQWP7ZHTXWPDJFANCNFSM6AAAAAA3Y5XHQU>
.
You are receiving this because you are subscribed to this thread.Message
ID: ***@***.***>
--
Sent from Gmail Mobile -Jared
|
Change this block of code:
For this:
With this you can use quantized models, CodeLlama, etc... |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Cool project!!!
There are a lot of models out there that will probably perform way better than vanilla llama 2. To get an idea, have a look at this HF space: https://huggingface.co/spaces/mike-ravkine/can-ai-code-results
Also, using a locally run 13b model is relatively trivial at this point, even with modest hardware if you use a quantized version. I am looking your code base to see where a good entry point for something like llama.cpp or textgen-webui could be harnessed.
The text was updated successfully, but these errors were encountered: