issues Search Results · repo:abacusai/gh200-llm language:Python
Filter by
11 results
(48 ms)11 results
inabacusai/gh200-llm (press backspace or delete to remove)vLLM now uses LMCache for KV caching and offloading. Can you include that into the docker build?
khayamgondal
- Opened 3 days ago
- #12
I m running it using apptainer.
$ apptainer pull docker://ghcr.io/abacusai/gh200-llm/llm-train-serve:latest
$ apptainer run apptainer run --nv --bind /projects llm-train-serve_latest.sif /bin/bash --login ...
zkx06111
- 1
- Opened 25 days ago
- #11
Hello and thank you for your great image!
I am a committer on the vLLM project and wanted to ask if you would be interested in contributing an official gh200
dockerfile to vLLM. The current dockerfile ...
mgoin
- Opened on Jan 29
- #10
Hi
I noticed that the latest release image is for Cuda 12.6. Do you also have an image for an older version where Cuda 12.4
is supported?
Best Manthan
manthan99
- Opened on Jan 17
- #9
Seeing following error when running train.sh data==0.4 installed Using your latest prebuilt docker image Launching
script via bash training/train.sh
[2024-09-26 18:05:30,229] [INFO] [launch.py:256:main] ...
khayamgondal
- Opened on Sep 26, 2024
- #7
Your Numbers are looking Promising. We experimented with the GH200 and vLLM too, but could not reproduce your high
numbers, so we are highly interested in the acutal scripts which produced this numbers. ...
Wetzr
- Opened on Sep 5, 2024
- #5
In the docker file you have references to custom built .whl files (e.g.
https://static.abacus.ai/pypi/abacusai/gh200-llm/pytorch-2407-cuda125/flash_attn-2.6.3-cp310-cp310-linux_aarch64.whl)
but there are ...
agm-eratosth
- Opened on Aug 28, 2024
- #4
Hi, @arvindsun
I m not sure if you ve seen this repository from the folks at LinkedIn, but I was wondering on the off chance you did if
you d had any luck getting this to work on GH200? I ve tried adding ...
jlotthammer
- 2
- Opened on Aug 24, 2024
- #3
The VLLM version used in dockerfile is 4.1 The docker container seems to use precompiled pypi packages. Please provide
an update for VLLM version 5.0 or insight into how to update this to 5.0
NiklasWilson
- 1
- Opened on Jun 12, 2024
- #2

Learn how you can use GitHub Issues to plan and track your work.
Save views for sprints, backlogs, teams, or releases. Rank, sort, and filter issues to suit the occasion. The possibilities are endless.Learn more about GitHub IssuesProTip!
Restrict your search to the title by using the in:title qualifier.
Learn how you can use GitHub Issues to plan and track your work.
Save views for sprints, backlogs, teams, or releases. Rank, sort, and filter issues to suit the occasion. The possibilities are endless.Learn more about GitHub IssuesProTip!
Press the /
key to activate the search input again and adjust your query.