[None][chore] Update flashinfer-python from 0.6.9 to 0.6.10#13746
[None][chore] Update flashinfer-python from 0.6.9 to 0.6.10#13746yihwang-nv merged 1 commit intoNVIDIA:mainfrom
Conversation
|
/bot run --disable-fail-fast |
📝 WalkthroughWalkthroughThis PR bumps ChangesFlashInfer Dependency Update
Estimated code review effort🎯 2 (Simple) | ⏱️ ~10 minutes Possibly related PRs
Suggested reviewers
🚥 Pre-merge checks | ✅ 4 | ❌ 1❌ Failed checks (1 warning)
✅ Passed checks (4 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches🧪 Generate unit tests (beta)
Tip 💬 Introducing Slack Agent: The best way for teams to turn conversations into code.Slack Agent is built on CodeRabbit's deep understanding of your code, so your team can collaborate across the entire SDLC without losing context.
Built for teams:
One agent for your entire SDLC. Right inside Slack. Comment |
|
PR_Github #46727 [ run ] triggered by Bot. Commit: |
Bump flashinfer-python dependency to the latest stable release. - Update version pins in requirements.txt, security_scanning/pyproject.toml, security_scanning/poetry.lock, and ATTRIBUTIONS-Python.md. - Replace string version comparison with packaging.version.Version in speculative/interface.py so the >= 0.6.4 gate evaluates correctly for versions like 0.6.10 (lexicographic compare would otherwise return False). Signed-off-by: Yihan Wang <yihwang@nvidia.com>
d199a79 to
da4929c
Compare
|
/bot run |
|
/bot run --disable-fail-fast |
|
PR_Github #46769 [ run ] triggered by Bot. Commit: |
|
PR_Github #46770 [ run ] triggered by Bot. Commit: |
|
PR_Github #46769 [ run ] completed with state |
|
PR_Github #46727 [ run ] completed with state |
|
PR_Github #46770 [ run ] completed with state |
juney-nvidia
left a comment
There was a problem hiding this comment.
Approved from the perspective of spec decoding related changes.
Summary
requirements.txt,security_scanning/pyproject.toml,security_scanning/poetry.lock, andATTRIBUTIONS-Python.md.packaging.version.Versionintensorrt_llm/_torch/speculative/interface.pyso the>= 0.6.4gate evaluates correctly. Lexicographic compare gave"0.6.10" >= "0.6.4"→ False, which would have silently disabled flashinfer inSpecWorkerBaseafter the bump.Test plan
pip install -r requirements.txtinstalls successfullypytest tests/unittest/_torch/flashinfer/ -vpytest tests/unittest/_torch/attention/test_flashinfer_attention.py -vSummary by CodeRabbit
Chores
Bug Fixes