forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 38
Insights: ROCm/vllm
Overview
Loading
Could not load contribution data
Please try again later
Loading
1 Release published by 1 person
-
v0.8.2+rocm
published
Mar 26, 2025
3 Pull requests merged by 1 person
-
Using torch commit that supports running scaled_mm on Radeon
#492 merged
Mar 25, 2025 -
Using temp dir for the intermediate P3L file
#491 merged
Mar 25, 2025 -
Upstream merge 2025 03 24
#489 merged
Mar 24, 2025
4 Pull requests opened by 4 people
-
WIP: Fixes to kernel tests
#487 opened
Mar 24, 2025 -
creating 1 gpu agents on OCI cluster
#488 opened
Mar 24, 2025 -
Updated README with March 25 Docker results
#490 opened
Mar 25, 2025 -
Docs_update_20250327
#493 opened
Mar 27, 2025
1 Issue closed by 2 people
-
[Bug]: Running Llama-2-70b inference on MI300x getting OOM
#397 closed
Mar 21, 2025
1 Issue opened by 1 person
-
[Bug]: Issue compiling Dockerfile.rocm_base
#486 opened
Mar 22, 2025
1 Unresolved conversation
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
Handling input dim size greater than 3 in tuned_gemm.py
#482 commented on
Mar 21, 2025 • 0 new comments