Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Wingman v2.0.8 , Local model is not supporting #27

Closed
NK-Aero opened this issue Dec 4, 2023 · 4 comments
Closed

Wingman v2.0.8 , Local model is not supporting #27

NK-Aero opened this issue Dec 4, 2023 · 4 comments

Comments

@NK-Aero
Copy link

NK-Aero commented Dec 4, 2023

Wingman v2.0.8 , Local model is not supporting. LM Studio, Koboldcpp URL are not working, please provide documentation or tutorial or video about enabling local models. Thank you

@nvms
Copy link
Owner

nvms commented Dec 4, 2023

I'm unfamiliar with LM Studio, but I just downloaded it and started a local inference server running llama 2 chat 7B q4_0 ggml. Using the OpenAI provider in Wingman, I changed the URL to http://localhost:1234/v1/chat/completions and things seem to just work (with the exception of recognizing the end of the completion stream, which may just be unique to LM Studio, but a bug none-the-less).

Can you provide some more information regarding your setup?

KoboldCpp support is almost finished - still porting this functionality from the previous major version.

Screenshot 2023-12-04 at 2 01 28 PM

@NK-Aero
Copy link
Author

NK-Aero commented Dec 4, 2023

Thank you , it is working now. i was using "http://localhost:1234/v1". now i changed to "http://localhost:1234/v1/chat/completions" . it is working.

I was using Wingman 1.3.8 preview, it was excellent. Now, it upgraded well. Thank you.

@nvms
Copy link
Owner

nvms commented Dec 4, 2023

with the exception of recognizing the end of the completion stream, which may just be unique to LM Studio, but a bug none-the-less

This is fixed now and the completion stream resoled by LM Studio should correctly end the response. Pushing a release with this bug fix now.

@nvms
Copy link
Owner

nvms commented Dec 4, 2023

Thank you , it is working now. i was using "http://localhost:1234/v1". now i changed to "http://localhost:1234/v1/chat/completions" . it is working.

I was using Wingman 1.3.8 preview, it was excellent. Now, it upgraded well. Thank you.

Excellent! Glad it's working. Have fun!

@nvms nvms closed this as completed Dec 4, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants