-
Notifications
You must be signed in to change notification settings - Fork 55
LCORE-336: bump-up Llama Stack to 0.2.17 #355
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LCORE-336: bump-up Llama Stack to 0.2.17 #355
Conversation
WalkthroughThe changes include updating the Changes
Estimated code review effort🎯 1 (Trivial) | ⏱️ ~3 minutes Possibly related PRs
Poem
Note ⚡️ Unit Test Generation is now available in beta!Learn more here, or try it out under "Finishing Touches" below. 📜 Recent review detailsConfiguration used: CodeRabbit UI 📒 Files selected for processing (1)
🚧 Files skipped from review as they are similar to previous changes (1)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (1)
✨ Finishing Touches🧪 Generate unit tests
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. 🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
SupportNeed help? Create a ticket on our support page for assistance with any issues or questions. CodeRabbit Commands (Invoked using PR comments)
Other keywords and placeholders
CodeRabbit Configuration File (
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 1
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
⛔ Files ignored due to path filters (1)
uv.lockis excluded by!**/*.lock
📒 Files selected for processing (1)
pyproject.toml(1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (1)
- GitHub Check: build-pr
| "llama-stack==0.2.17", | ||
| "llama-stack-client==0.2.17", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💡 Verification agent
🧩 Analysis chain
Verify 0.2.17 exists & still supports Python 3.12, and consider relaxing the hard pin
Good to see both libs kept in sync, but two quick checks would de-risk the bump:
- Make sure
llama-stack==0.2.17andllama-stack-client==0.2.17are published on PyPI and list Python 3.12 in their classifiers. - Hard-pinning (
==) stops you from receiving future patch releases (0.2.18, 0.2.19…) that may deliver security fixes. A more typical spec is “at least this version but stay on the 0.2 line”.
Example diff:
- "llama-stack==0.2.17",
- "llama-stack-client==0.2.17",
+ "llama-stack>=0.2.17,<0.3",
+ "llama-stack-client>=0.2.17,<0.3",Quick verification script:
🏁 Script executed:
#!/usr/bin/env bash
set -euo pipefail
for pkg in llama-stack llama-stack-client; do
echo "Checking $pkg 0.2.17 on PyPI…"
curl -s "https://pypi.org/pypi/${pkg}/json" | jq -r --arg v "0.2.17" \
'.releases[$v] | if . == null then "NOT FOUND" else "Found" end'
doneLength of output: 580
Let’s verify that both packages declare Python 3.12 in their PyPI classifiers:
🏁 Script executed:
#!/usr/bin/env bash
set -euo pipefail
for pkg in llama-stack llama-stack-client; do
echo "Checking Python 3.12 support for $pkg 0.2.17…"
curl -s "https://pypi.org/pypi/${pkg}/json" \
| jq -r '.info.classifiers[]' \
| grep -q "Programming Language :: Python :: 3.12" \
&& echo "✔️ $pkg supports Python 3.12" \
|| echo "❌ $pkg does NOT list Python 3.12"
doneLength of output: 802
Address Python 3.12 support in llama-stack & relax hard pins
Both packages are published at 0.2.17, but the llama-stack release doesn’t list Python 3.12 as supported, which may block installations on 3.12 environments:
• In pyproject.toml (lines 28–29):
– “llama-stack==0.2.17” exists but ❌ lacks a 3.12 classifier
– “llama-stack-client==0.2.17” ✅ supports Python 3.12
Please either ensure llama-stack 0.2.17 is updated to declare Python 3.12 support (or bump to a version that does) and consider loosening the pins so you continue to get patch releases:
- "llama-stack==0.2.17",
- "llama-stack-client==0.2.17",
+ "llama-stack>=0.2.17,<0.3",
+ "llama-stack-client>=0.2.17,<0.3",📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
| "llama-stack==0.2.17", | |
| "llama-stack-client==0.2.17", | |
| "llama-stack>=0.2.17,<0.3", | |
| "llama-stack-client>=0.2.17,<0.3", |
🤖 Prompt for AI Agents
In pyproject.toml around lines 28 to 29, the dependency "llama-stack==0.2.17"
lacks declared support for Python 3.12, which may cause installation issues. To
fix this, verify if a newer version of "llama-stack" exists that includes Python
3.12 support and update the version accordingly. Also, relax the version pinning
by changing from exact "==" pins to a compatible range (e.g., "~=") to allow
patch updates while ensuring compatibility.
umago
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Fetched the PR locally, did some manual checks (query/streaming_query) and it seems good. CI upstream also green so LGTM
|
Don't forget |
|
/hold Suspected llama-stack regression |
|
/unhold Nevermind, it's a bit complicated but tl;dr it's because in assisted-chat we did: - provider_id: gemini
provider_type: remote::gemini
config:
api_key: ${env.GEMINI_API_KEY:+}instead of - provider_id: gemini
provider_type: remote::gemini
config:
api_key: ${env.GEMINI_API_KEY:=}(we copied the llama-stack config from their repo which they since fixed) (it worked perfectly fine because litellm picked our keys regardless, until they added an explicit error in llama-stack) |
Description
LCORE-336: bump-up Llama Stack to 0.2.17
Type of change
Related Tickets & Documents
Summary by CodeRabbit
llama-stackandllama-stack-client.