Conversation
|
MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅ |
Summary of ChangesHello @arekay-nv, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request introduces a new example for benchmarking the Highlights
🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console. Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
8a2d1d2 to
a750ece
Compare
There was a problem hiding this comment.
Code Review
This pull request introduces a new example configuration for benchmarking gpt-oss-120b. My review identifies a couple of issues in the new YAML configuration file. The specified benchmark durations are extremely short and inconsistent with the comments, which could lead to skewed or meaningless results. Additionally, the report directory name is hardcoded with a specific date, which reduces the reusability of this example. I've provided suggestions to align the durations with their descriptions and to make the directory name more generic.
There was a problem hiding this comment.
Pull request overview
This PR adds a new example configuration for benchmarking the GPT-OSS-120B model, providing a reference implementation for LLM working group proposals. The example includes performance testing setup with concurrent load patterns and comprehensive metric collection.
Key Changes:
- Added
gpt-oss-120b-benchmarkYAML configuration with model parameters and runtime settings - Configured performance benchmarking with 512 concurrent connections targeting the 120B parameter model
- Set up metric collection for throughput, latency, TTFT, and TPOT measurements
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
a750ece to
b2220c5
Compare
b2220c5 to
f173b89
Compare
There was a problem hiding this comment.
Pull request overview
Copilot reviewed 2 out of 2 changed files in this pull request and generated 3 comments.
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
f173b89 to
1895a17
Compare
Signed-off-by: Rashid Kaleem <230885705+arekay-nv@users.noreply.github.com>
1895a17 to
879674f
Compare
There was a problem hiding this comment.
Pull request overview
Copilot reviewed 3 out of 3 changed files in this pull request and generated 1 comment.
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
There was a problem hiding this comment.
Pull request overview
Copilot reviewed 3 out of 3 changed files in this pull request and generated 3 comments.
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
What does this PR do?
Adds a
gpt-oss-120bexample compatible with the LLM working groups proposal.Performance and accuracy datasets added via
git-lfsType of change
Related issues
Testing
Checklist