Replies: 1 comment
-
You can independently launch llama2 on each node. Then manage them with k8s. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I want to use 2 node(2*8GPU) to accelerate the inference speed of the llama2.
Beta Was this translation helpful? Give feedback.
All reactions