-
I'm trying to build a docker image with any model. However, after succesfull Bento build the containerize command fails with this error: $ bentoml containerize facebook-opt-1-3b-service:8c7b10754972749675d22364c25c428b29face51
Building OCI-compliant image for facebook-opt-1-3b-service:8c7b10754972749675d22364c25c428b29face51 with docker
[+] Building 2.3s (3/3) FINISHED docker:default
=> [internal] load build definition from Dockerfile 0.1s
=> => transferring dockerfile: 3.36kB 0.0s
=> [internal] load .dockerignore 0.2s
=> => transferring context: 2B 0.0s
=> ERROR [internal] load metadata for docker.io/nvidia/cuda:11.7.0-cudnn8-runtime-ubuntu20.04 1.9s
------
> [internal] load metadata for docker.io/nvidia/cuda:11.7.0-cudnn8-runtime-ubuntu20.04:
------
Dockerfile:6
--------------------
4 |
5 | # Block SETUP_BENTO_BASE_IMAGE
6 | >>> FROM nvidia/cuda:11.7.0-cudnn8-runtime-ubuntu20.04 as base-container
7 |
8 | ENV LANG=C.UTF-8
--------------------
ERROR: failed to solve: nvidia/cuda:11.7.0-cudnn8-runtime-ubuntu20.04: docker.io/nvidia/cuda:11.7.0-cudnn8-runtime-ubuntu20.04: not found The docker tag does not exist. I saw in the documentation that it is possible to set the docker:
cuda_version: "11.7.1" But after adding the |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 5 replies
-
I haven't released a new version yet, but you can try out the main build pip install "git+https://github.com/bentoml/openllm" I think the responsibility of CUDA should be at OpenLLM |
Beta Was this translation helpful? Give feedback.
I haven't released a new version yet, but you can try out the main build
I think the responsibility of CUDA should be at OpenLLM