-
-
Notifications
You must be signed in to change notification settings - Fork 4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEAT]: Support Ktransformers as LLM #3363
Comments
If this is just another inference service that is OpenAI Compatible, you can just use the base URL and connection information in the Generic OpenAI Connector for the LLM and it will work the same. |
I am currently trying to run KTransformer (Deepseek-R1) + Generic OpenAI Connector with AnythingLLM on an Apple M1 computer. While the server shows normal responses, the client keeps getting stuck at the last returned character until I manually click the pause button, after which the final character is displayed. |
Sounds like KTransformers does not return a stop reason - thus why the connection stream does not close. Can you confirm if the last chunk from KTransformer returns a stop reason? |
does it work? |
That should work anything-llm/server/utils/AiProviders/genericOpenAi/index.js Lines 312 to 316 in fab7403
|
What would you like to see?
When will AnythingLLmm support ktransformers, the ktransformers can let the AI use less hardware to complete high quantity answers!!!
Here is the link:
https://github.com/kvcache-ai/ktransformers
The text was updated successfully, but these errors were encountered: