Skip to content

llm_sparsity: Set warmup_steps 0 instead of 0.0 for transformers 5.x compat#1393

Merged
kaix-nv merged 1 commit into
mainfrom
kevalmorabia97-patch-1
May 5, 2026
Merged

llm_sparsity: Set warmup_steps 0 instead of 0.0 for transformers 5.x compat#1393
kaix-nv merged 1 commit into
mainfrom
kevalmorabia97-patch-1

Conversation

@kevalmorabia97
Copy link
Copy Markdown
Collaborator

@kevalmorabia97 kevalmorabia97 commented May 5, 2026

Fix for NVBug 6120631 to fix

finetune.py: error: argument --warmup_steps/--warmup-steps: invalid int value: '0.0'

Summary by CodeRabbit

  • Chores
    • Corrected parameter format in finetuning example script for consistency.

…compat

Signed-off-by: Keval Morabia <28916987+kevalmorabia97@users.noreply.github.com>
@kevalmorabia97 kevalmorabia97 requested a review from a team as a code owner May 5, 2026 13:13
@kevalmorabia97 kevalmorabia97 added the cherry-pick-0.44.0 After code freeze, cherry-pick to release branch for next rc (bulk update). Only for bug fixes / doc label May 5, 2026
@kevalmorabia97 kevalmorabia97 requested a review from realAsma May 5, 2026 13:13
@coderabbitai
Copy link
Copy Markdown
Contributor

coderabbitai Bot commented May 5, 2026

No actionable comments were generated in the recent review. 🎉

ℹ️ Recent review info
⚙️ Run configuration

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Enterprise

Run ID: a6897c5e-a746-45aa-a0b8-d0dfd229ef79

📥 Commits

Reviewing files that changed from the base of the PR and between d1ed76d and fead988.

📒 Files selected for processing (1)
  • examples/llm_sparsity/weight_sparsity/launch_finetune.sh

📝 Walkthrough

Walkthrough

A shell script for launching model finetuning updates the --warmup_steps argument from 0.0 (float) to 0 (integer) in the accelerate launch command.

Changes

Script Configuration

Layer / File(s) Summary
Launch Script
examples/llm_sparsity/weight_sparsity/launch_finetune.sh
--warmup_steps argument changed from 0.0 to 0 in the accelerate launch command.

Estimated code review effort

🎯 1 (Trivial) | ⏱️ ~2 minutes

🚥 Pre-merge checks | ✅ 6
✅ Passed checks (6 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title accurately summarizes the main change: updating warmup_steps from 0.0 to 0 for transformers 5.x compatibility.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
Linked Issues check ✅ Passed Check skipped because no linked issues were found for this pull request.
Out of Scope Changes check ✅ Passed Check skipped because no linked issues were found for this pull request.
Security Anti-Patterns ✅ Passed PR modifies only a shell script, not Python code. Security check requires "Python changes" which this PR lacks. Check not applicable.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
📝 Generate docstrings
  • Create stacked PR
  • Commit on current branch
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Commit unit tests in branch kevalmorabia97-patch-1

Comment @coderabbitai help to get the list of available commands and usage tips.

@kevalmorabia97 kevalmorabia97 requested review from kaix-nv and rohansjoshi and removed request for realAsma May 5, 2026 13:15
@codecov
Copy link
Copy Markdown

codecov Bot commented May 5, 2026

Codecov Report

✅ All modified and coverable lines are covered by tests.
✅ Project coverage is 77.07%. Comparing base (f0eaa19) to head (fead988).
⚠️ Report is 1 commits behind head on main.

Additional details and impacted files
@@            Coverage Diff             @@
##             main    #1393      +/-   ##
==========================================
+ Coverage   75.74%   77.07%   +1.32%     
==========================================
  Files         476      476              
  Lines       51057    51057              
==========================================
+ Hits        38672    39350     +678     
+ Misses      12385    11707     -678     
Flag Coverage Δ
examples 41.56% <ø> (+1.83%) ⬆️
unit 52.31% <ø> (ø)

Flags with carried forward coverage won't be shown. Click here to find out more.

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

@kaix-nv kaix-nv merged commit d794595 into main May 5, 2026
38 checks passed
@kaix-nv kaix-nv deleted the kevalmorabia97-patch-1 branch May 5, 2026 17:06
@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented May 5, 2026

PR Preview Action v1.8.1
Preview removed because the pull request was closed.
2026-05-05 17:06 UTC

@kevalmorabia97 kevalmorabia97 added the cherry-pick-done Added by bot once PR is cherry-picked to the release branch label May 11, 2026
kevalmorabia97 added a commit that referenced this pull request May 11, 2026
#1416 (#1426)

## Cherry-picked PRs

- #1393
- #1389
- #1268
- #1397
- #1402
- #1411
- #1410
- #1419
- #1408
- #1416

<!-- This is an auto-generated comment: release notes by coderabbit.ai
-->
## Summary by CodeRabbit

* **New Features**
* SPEEDBench now uses stratified sampling for deterministic, balanced
dataset selection.
* Added legacy quantization conversion shims for INT4, MXFP8 and FP4→2DQ
workflows.
* AWQ Lite: fallback handling for uncalibrated per-expert quantizers
during export.

* **Bug Fixes**
  * Clamp FP8 scales in NVFP4 quantization to avoid NaNs.
  * Fixed warmup steps formatting in finetune launch script.

* **Improvements**
  * LM-Eval integration updated for v0.4.10+ compatibility.
  * TensorRT execution routed through a dedicated trtexec helper.

* **Tests**
* Added/regressed tests covering quantization shims, FP8 scale behavior,
export fallbacks, and LM eval.

[![Review Change
Stack](https://storage.googleapis.com/coderabbit_public_assets/review-stack-in-coderabbit-ui.svg)](https://app.coderabbit.ai/change-stack/NVIDIA/Model-Optimizer/pull/1426)
<!-- end of auto-generated comment: release notes by coderabbit.ai -->

---------

Signed-off-by: Keval Morabia <28916987+kevalmorabia97@users.noreply.github.com>
Signed-off-by: Alexandre Milesi <milesial@users.noreply.github.com>
Signed-off-by: gcunhase <4861122+gcunhase@users.noreply.github.com>
Signed-off-by: Chenjie Luo <chenjiel@nvidia.com>
Signed-off-by: weimingc <17592131+meenchen@users.noreply.github.com>
Signed-off-by: weimingc <weimingc@nvidia.com>
Signed-off-by: Suguna Velury <178320438+sugunav14@users.noreply.github.com>
Signed-off-by: ajrasane <131806219+ajrasane@users.noreply.github.com>
Co-authored-by: milesial <milesial@users.noreply.github.com>
Co-authored-by: Gwena Cunha <4861122+gcunhase@users.noreply.github.com>
Co-authored-by: Chenjie Luo <108829653+cjluo-nv@users.noreply.github.com>
Co-authored-by: Wei-Ming Chen <17592131+meenchen@users.noreply.github.com>
Co-authored-by: sugunav14 <178320438+sugunav14@users.noreply.github.com>
Co-authored-by: Ajinkya Rasane <131806219+ajrasane@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

cherry-pick-0.44.0 After code freeze, cherry-pick to release branch for next rc (bulk update). Only for bug fixes / doc cherry-pick-done Added by bot once PR is cherry-picked to the release branch

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants