New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bug: openllm start opt or openllm start dolly-v2 faild #47
Comments
openllm start dolly-v2 is OK as below |
http://localhost:3000/readyz |
are you seeing this with dolly, or with both? I have tested OPT on my end, on both linux and macos and it startup just fine |
Hey, I have fixed this issue on main and will release a patch version soon. |
OK, THANKS. |
Hey @aarnphm, I ran into the exact same behavior. |
I tried it on the system (RHEL 8.4) outside of the container with a venv (Python 3.9), the readyz endpoint also indicates |
Can you dumped the whole stack trace in a new issue? |
Is there a commit or a branch where you can see these changes? I did not find any resources. |
Containerizing Bento with podman should already be supported. bentoml containerize llm-bento --backend podman --opt ... |
Tho there is an internal bug that I just discovered recently wrt running within the container. I will post updates about this soon |
I believe this is related to the container deployment. can you create a new issue? thanks |
Describe the bug
openllm start opt and openllm start dolly-v2 shows OK.
when i made the query below came out.
2023-06-21T16:45:40+0800 [INFO] [runner:llm-dolly-v2-runner:1] _ (scheme=http,method=GET,path=http://127.0.0.1:8000/readyz,type=,length=) (status=404,type=text/plain; charset=utf-8,length=9) 0.574ms (trace=100fa96d33433a772259d444a0006ca9,span=4caf6df55eb0c67e,sampled=1,service.name=llm-dolly-v2-runner)
2023-06-21T16:45:40+0800 [INFO] [api_server:llm-dolly-v2-service:9] 127.0.0.1:63613 (scheme=http,method=GET,path=/readyz,type=,length=) (status=503,type=text/plain; charset=utf-8,length=22) 5.220ms (trace=100fa96d33433a772259d444a0006ca9,span=7ec016176efc036d,sampled=1,service.name=llm-dolly-v2-service)
To reproduce
No response
Logs
Environment
Environment variable
System information
bentoml
: 1.0.22python
: 3.8.16platform
: macOS-13.4-arm64-arm-64bituid_gid
: 501:20conda
: 23.3.1in_conda_env
: Trueconda_packages
pip_packages
transformers
version: 4.30.2System information (Optional)
apple m1 max
The text was updated successfully, but these errors were encountered: