-
Notifications
You must be signed in to change notification settings - Fork 6.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[chatglm] Server Unavailable Error, Internal Server Error #3066
Comments
Hey @zps-509, good to see you back! Running into a bit of a quirky issue, huh? Let’s see what we can do about that. Based on the information you've provided, it seems like the issue you're experiencing might be related to how the response from the stable diffusion process is being handled. The error message "[chatglm] Server Unavailable Error, Internal Server Error" suggests that there might be an issue with the server hosting the ChatGPT service, but it could also be a symptom of how the response is being processed. A similar issue was discussed in the Completion Request Error thread. The solution suggested there was to differentiate between simple and streaming responses in your client code. For simple JSON responses, you can access the response body directly. However, for streaming responses, you should consume the stream properly instead of trying to access it as plain text. Here's how you can handle both types of responses correctly:
If the issue persists, it might be worth checking the logs of each container to identify any errors. You can use the command As for the role of the chatglm3 model in the Dify setup and its interaction with the stable diffusion process, I wasn't able to find specific information within the repository. However, it's likely that the chatglm3 model is used to process and generate the conversational responses, while the stable diffusion process is used for image generation. The issue seems to arise when these two processes interact, possibly due to how the response from the stable diffusion process is being handled. I hope this helps! If you have any more information or if there's anything else you'd like me to look into, please let me know. Sources
|
我刚刚在 https://udify.app/chat/VfbjY8kvFH76I3CA 配置同一个 chatglm 和同一个 stable diffusion成功了。可以生成图片,没有报错。。 |
INFO:httpx:HTTP Request: GET http://192.168.133.98:20009/sdapi/v1/sd-models "HTTP/1.1 200 OK" During handling of the above exception, another exception occurred: Traceback (most recent call last): 这个是 docker_api 的日志,请帮忙看下怎么处理 |
[root@localhost docker]$ curl -X POST "http://192.168.133.98:55555/v1/chat/completions" -H "Content-Type: application/json" -d "{"model": "chatglm3-6b", "messages": [{"role": "system", "content": ""}, {"role": "user", "content": "hello Word"}], "stream": false, "max_tokens": 100, "temperature": 0.8, "top_p": 0.8}" {"model":"chatglm3-6b","object":"chat.completion","choices":[{"index":0,"message":{"role":"assistant","content":"Hello! How can I assist you today?","name":null,"function_call":null},"finish_reason":"stop"}],"created":1711979906,"usage":{"prompt_tokens":11,"total_tokens":23,"completion_tokens":12}} |
This is most likely because the official demo of OpenAI server in the ChatGLM repo is not a productive solution, as it always raises errors, for a better experience, we recommend XInference as the inferencing service. |
我切换成 XInference ,可以正常调取tools,没有报错。 |
Sounds weird, could you pls provide a bit more details about that? |
以上操作对比,感觉使用 XInference之后,提示词失效了。 |
Hi, @zps-509, I'm helping the Dify team manage their backlog and am marking this issue as stale. It looks like you encountered a "Server Unavailable Error, Internal Server Error" when using DIFY with chatglm3 and stable diffusion. There were some detailed responses from user dosubot suggesting potential solutions and troubleshooting steps, as well as a recommendation from user Yeuoly to use XInference as the inferencing service for a better experience. However, it seems that the issue remains unresolved despite the suggested actions. Could you please confirm if this issue is still relevant to the latest version of the Dify repository? If it is, please let the Dify team know by commenting on the issue. Otherwise, feel free to close the issue yourself, or it will be automatically closed in 7 days. Thank you! |
Please upgrade to latest version to see if this issue persists. :) |
Self Checks
Dify version
0.5.11
Cloud or Self Hosted
Self Hosted (Docker)
Steps to reproduce
✔️ Expected Behavior
No response
❌ Actual Behavior
No response
The text was updated successfully, but these errors were encountered: