Skip to content

Conversation

namanlalitnyu
Copy link
Contributor

@namanlalitnyu namanlalitnyu commented Aug 27, 2025

  • These changes upload the SGLang benchmark results to the AWS S3 bucket, which internally triggers the AWS Lambda function to upload them to the clickhouse database. Eventually, being rendered in the HUD Dashboard.

  • For SGLang benchmarking, currently, the vllm bench serve command was generating a test.pytorch.json file containing the Pytorch formatted benchmark results, but it was hardcoding the benchmark name to vllm benchmark. Due to that, whenever the SGLang benchmarking was being run, it was also allocating the same benchmark name. This was leading to issues in the HUD dashboard where the clickhouse sql queries were trying to find the benchmark name as SGLang benchmark, but due to this issue, the results were returning an empty array.
    So, added a check in the bash script to replace the benchmark name with correct one, as the original implementation is being done in the vllm code repository.

Testing:

Verified the changes from the Github workflow that the results are getting uploaded correctly to AWS S3

Screenshot 2025-08-27 at 3 43 55 PM

Verified the changes from the flambeau dashboard that the results are getting uploaded to Clickhouse database.

Screenshot 2025-08-27 at 3 55 59 PM Screenshot 2025-08-27 at 3 56 08 PM

@meta-cla meta-cla bot added the cla signed label Aug 27, 2025
@namanlalitnyu namanlalitnyu changed the title [SGLang Workflow] Upload benchmark results to S3 and Clickhouse [SGLang Workflow] Upload benchmark results to AWS S3 Aug 27, 2025
echo "⚠️ No benchmark results found in ${BENCHMARK_RESULTS}" >> $GITHUB_STEP_SUMMARY
fi
python3 .github/scripts/upload_benchmark_results.py \
Copy link
Contributor

@huydhn huydhn Aug 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I notice that this workflow doesn't have a schedule yet. Are you planning to add one? I'm thinking daily, biweekly, or weekly depending on how frequent we need to look at sglang results.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for pointing it out, I have added it for weekly as of now, as we discussed initially. SGLang releases the new stable versions every monthly, but their pre-release versions mostly happen every week, so I think it would be fine to have it weekly for now. But, in future, we can update it to lesser/more as well based on the feedback.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think weekly is a good starting point

"dataset_path": "./ShareGPT_V3_unfiltered_cleaned_split.json",
"num_prompts": 200
}
},
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Added a couple of tests for having more metrics on the SGLang dashboard.

Image

@namanlalitnyu namanlalitnyu merged commit d1940e4 into main Aug 29, 2025
3 of 13 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants