diff --git a/.github/workflows/build-and-deploy.yml b/.github/workflows/build-and-deploy.yml index 5c16741..ae90238 100644 --- a/.github/workflows/build-and-deploy.yml +++ b/.github/workflows/build-and-deploy.yml @@ -1,7 +1,6 @@ name: Build and Deploy Docusaurus Documentation on: - # Allow manual trigger workflow_dispatch: inputs: version: @@ -35,134 +34,85 @@ jobs: with: token: ${{ secrets.GITHUB_TOKEN }} - - name: Clone llama-stack repository with full history + - name: Clone llama-stack repository run: | - echo "๐Ÿ“ฅ Cloning llama-stack repository with full history..." - git clone https://github.com/llamastack/llama-stack.git ${{ runner.temp }}/llama-stack - cd ${{ runner.temp }}/llama-stack + echo "๐Ÿ“ฅ Cloning llama-stack repository..." + TEMP_DIR=$(mktemp -d) + echo "TEMP_DIR=$TEMP_DIR" >> $GITHUB_ENV + + git clone https://github.com/llamastack/llama-stack.git "$TEMP_DIR/llama-stack" + cd "$TEMP_DIR/llama-stack" git fetch --all --tags - # Determine version to build + # Determine version to checkout VERSION="${{ github.event.inputs.version }}" - if [ -z "$VERSION" ] || [ "$VERSION" = "latest" ]; then - # Use main branch for latest + if [ "$VERSION" = "latest" ] || [ -z "$VERSION" ]; then git checkout main - echo "Building from main branch (latest)" + echo "โœ… Using main branch (latest)" echo "BUILDING_LATEST=true" >> $GITHUB_ENV echo "VERSION_TAG=latest" >> $GITHUB_ENV else - # Checkout specific version tag git checkout "$VERSION" - echo "Building from tag: $VERSION" + echo "โœ… Using tag: $VERSION" echo "BUILDING_LATEST=false" >> $GITHUB_ENV echo "VERSION_TAG=$VERSION" >> $GITHUB_ENV fi - echo "โœ… Repository prepared" - - name: Set up Node.js uses: actions/setup-node@v4 with: node-version: '20' - cache: 'npm' - cache-dependency-path: '${{ runner.temp }}/llama-stack/docs/package-lock.json' - - name: Install Docusaurus dependencies + - name: Install dependencies and setup versioning run: | - echo "๐Ÿ“ฆ Installing Docusaurus dependencies..." - cd ${{ runner.temp }}/llama-stack/docs + echo "๐Ÿ“ฆ Installing dependencies..." + cd "${{ env.TEMP_DIR }}/llama-stack/docs" npm ci echo "โœ… Dependencies installed" - - name: Apply Docusaurus configuration patches - run: | - echo "โš™๏ธ Applying Docusaurus configuration patches..." - cd ${{ runner.temp }}/llama-stack/docs + echo "โš™๏ธ Setting up versioning configuration..." - # Create versioning files for current state - cat > versionsArchived.json << 'EOF' - { - "v0.2.22": "https://llamastack.github.io/legacy/v0.2.22/", - "v0.2.21": "https://llamastack.github.io/legacy/v0.2.21/", - "v0.2.20": "https://llamastack.github.io/legacy/v0.2.20/", - "v0.2.19": "https://llamastack.github.io/legacy/v0.2.19/", - "v0.2.18": "https://llamastack.github.io/legacy/v0.2.18/", - "v0.2.17": "https://llamastack.github.io/legacy/v0.2.17/", - "v0.2.16": "https://llamastack.github.io/legacy/v0.2.16/", - "v0.2.15": "https://llamastack.github.io/legacy/v0.2.15/", - "v0.2.14": "https://llamastack.github.io/legacy/v0.2.14/", - "v0.2.13": "https://llamastack.github.io/legacy/v0.2.13/", - "v0.2.12": "https://llamastack.github.io/legacy/v0.2.12/", - "v0.2.11": "https://llamastack.github.io/legacy/v0.2.11/" - } - EOF + # Copy persistent versionsArchived.json from repo + cp "${{ github.workspace }}/versionsArchived.json" ./ - # Load current Docusaurus versions + # Load existing versions.json from deployed site or create empty if [ -f "${{ github.workspace }}/docs/versions.json" ]; then cp "${{ github.workspace }}/docs/versions.json" ./ + echo "โœ… Loaded existing versions.json" else echo "[]" > versions.json + echo "โœ… Created empty versions.json" fi - echo "๐Ÿ”ง Patching Docusaurus configuration..." - - # Create comprehensive config patch script - cat > patch-config.js << 'PATCH_EOF' + # Patch Docusaurus config for versioning + node << 'EOF' const fs = require('fs'); - const path = require('path'); - - const configPath = 'docusaurus.config.ts'; - let config = fs.readFileSync(configPath, 'utf8'); - // Load versions to determine current version label - const buildingLatest = process.env.BUILDING_LATEST === 'true'; - const versionTag = process.env.VERSION_TAG; - const currentVersionLabel = buildingLatest ? 'Latest' : versionTag; + let config = fs.readFileSync('docusaurus.config.ts', 'utf8'); - console.log(`Patching config for version: ${versionTag} (latest: ${buildingLatest})`); - - // Add versioning imports and logic + // Add versioning imports after OpenAPI import const versioningImports = ` - // Import archived versions configuration + // Import fs for versioning configuration const fs = require('fs'); - const path = require('path'); - // Load archived versions (legacy Sphinx versions) + // Versioning configuration for llamastack.github.io const versionsArchived = (() => { try { - return JSON.parse(fs.readFileSync(path.join(__dirname, 'versionsArchived.json'), 'utf8')); + return JSON.parse(fs.readFileSync('./versionsArchived.json', 'utf8')); } catch (e) { console.warn('Could not load versionsArchived.json:', e); return {}; } })(); - // Load current Docusaurus versions - const currentVersions = (() => { - try { - return JSON.parse(fs.readFileSync(path.join(__dirname, 'versions.json'), 'utf8')); - } catch (e) { - console.warn('Could not load versions.json:', e); - return []; - } - })(); - - // Create dropdown items for archived versions (legacy Sphinx) const archivedVersionsDropdownItems = Object.entries(versionsArchived).map( ([versionName, versionUrl]) => ({ label: versionName, href: versionUrl, }) ); - - // Create dropdown items for Docusaurus versions - const docusaurusVersionsDropdownItems = currentVersions.map(version => ({ - label: version, - to: \`/docs/\${version}/\`, - })); `; - // Insert versioning imports after existing imports config = config.replace( /import type \* as OpenApiPlugin from "docusaurus-plugin-openapi-docs";/, `import type * as OpenApiPlugin from "docusaurus-plugin-openapi-docs"; @@ -170,64 +120,26 @@ jobs: ${versioningImports}` ); - // Update version configuration based on build type - const versionConfig = buildingLatest ? - `// Versioning configuration - lastVersion: 'current', - versions: { - current: { - label: '${currentVersionLabel}', - path: '', - }, - }, - onlyIncludeVersions: ['current'],` : - `// Versioning configuration - lastVersion: 'current', - versions: { - current: { - label: '${currentVersionLabel}', - path: '', - }, - }, - includeCurrentVersion: true,`; - - // Add/update versioning configuration to docs config - if (config.includes('docItemComponent: "@theme/ApiItem"')) { - config = config.replace( - /docItemComponent: "@theme\/ApiItem", \/\/ Derived from docusaurus-theme-openapi/, - `docItemComponent: "@theme/ApiItem", // Derived from docusaurus-theme-openapi - - ${versionConfig}` - ); - } - - // Create version dropdown combining Docusaurus and archived versions + // Add version dropdown to navbar (replace GitHub item) const versionDropdown = ` { href: 'https://github.com/llamastack/llama-stack', label: 'GitHub', position: 'right', }, - // Version dropdown with current, Docusaurus, and archived versions { type: 'docsVersionDropdown', position: 'right', - dropdownItemsAfter: [ - // Docusaurus versions (if any) - ...docusaurusVersionsDropdownItems, - // Separator before archived versions - ...(archivedVersionsDropdownItems.length > 0 ? [ - { - type: 'html', - value: '', - }, - { - type: 'html', - className: 'dropdown-archived-versions', - value: 'Archived versions', - }, - ...archivedVersionsDropdownItems, - ] : []), - // All versions link + dropdownItemsAfter: archivedVersionsDropdownItems.length > 0 ? [ + { + type: 'html', + value: '', + }, + { + type: 'html', + className: 'dropdown-archived-versions', + value: 'Archived versions', + }, + ...archivedVersionsDropdownItems, { type: 'html', value: '', @@ -236,151 +148,183 @@ jobs: to: '/versions', label: 'All versions', }, - ], + ] : [], },`; - // Replace GitHub item with version dropdown + GitHub config = config.replace( - / {\s*href: 'https:\/\/github\.com\/llamastack\/llama-stack',\s*label: 'GitHub',\s*position: 'right',\s*},/, + /\s*{\s*href:\s*'https:\/\/github\.com\/llamastack\/llama-stack',\s*label:\s*'GitHub',\s*position:\s*'right',\s*},/, versionDropdown ); - fs.writeFileSync(configPath, config); - console.log('โœ… Docusaurus configuration patched successfully'); - PATCH_EOF - - # Apply the patch - BUILDING_LATEST=${{ env.BUILDING_LATEST }} VERSION_TAG=${{ env.VERSION_TAG }} node patch-config.js + fs.writeFileSync('docusaurus.config.ts', config); + console.log('โœ… Configuration patched'); + EOF echo "โœ… Configuration patches applied" - - name: Create Docusaurus version (if not latest) - if: env.BUILDING_LATEST != 'true' + - name: Import existing versioning artifacts run: | - echo "๐Ÿ“š Creating Docusaurus version for ${{ env.VERSION_TAG }}..." - cd ${{ runner.temp }}/llama-stack/docs + echo "๐Ÿ“ฅ Importing existing versioning artifacts from repository..." + cd "${{ env.TEMP_DIR }}/llama-stack/docs" - # Generate API docs first - npm run gen-api-docs all + # Copy existing versioned_docs if they exist + if [ -d "${{ github.workspace }}/versioned_docs" ]; then + cp -r "${{ github.workspace }}/versioned_docs" ./ + echo "โœ… Imported existing versioned_docs" + else + echo "โ„น๏ธ No existing versioned_docs found (first version)" + fi - # Create the version - npm run docusaurus docs:version ${{ env.VERSION_TAG }} + # Copy existing versioned_sidebars if they exist + if [ -d "${{ github.workspace }}/versioned_sidebars" ]; then + cp -r "${{ github.workspace }}/versioned_sidebars" ./ + echo "โœ… Imported existing versioned_sidebars" + else + echo "โ„น๏ธ No existing versioned_sidebars found (first version)" + fi + + # Copy existing versions.json if it exists (overrides what we loaded earlier) + if [ -f "${{ github.workspace }}/versions.json" ]; then + cp "${{ github.workspace }}/versions.json" ./ + echo "โœ… Imported existing versions.json" + else + echo "โ„น๏ธ No existing versions.json found (first version)" + fi - echo "โœ… Docusaurus version ${{ env.VERSION_TAG }} created" + echo "โœ… Versioning artifacts import completed" - - name: Generate API documentation + - name: Build documentation run: | - echo "๐Ÿ“š Generating API documentation..." - cd ${{ runner.temp }}/llama-stack/docs + cd "${{ env.TEMP_DIR }}/llama-stack/docs" + + # Generate API docs first (required for current build) npm run gen-api-docs all echo "โœ… API docs generated" + # Create version if not latest (after content is ready) + if [ "${{ env.BUILDING_LATEST }}" != "true" ]; then + echo "๐Ÿ“š Creating Docusaurus version: ${{ env.VERSION_TAG }}" + + # Create the version snapshot + npx docusaurus docs:version "${{ env.VERSION_TAG }}" + + # Ensure prompt-format.png is available where versioned docs expect it + mkdir -p versioned_docs/resources + if [ ! -f "versioned_docs/resources/prompt-format.png" ]; then + echo "๐Ÿ“ฅ Downloading missing prompt-format.png..." + curl -o versioned_docs/resources/prompt-format.png https://raw.githubusercontent.com/llamastack/llama-stack/main/docs/static/img/prompt-format.png + echo "โœ… Downloaded prompt-format.png to versioned_docs/resources/" + else + echo "โœ… prompt-format.png already exists in versioned_docs/resources/" + fi + + echo "โœ… Version ${{ env.VERSION_TAG }} created" + else + echo "๐Ÿ—๏ธ Building latest version (no version snapshot needed)" + fi + - name: Build Docusaurus site run: | echo "๐Ÿ—๏ธ Building Docusaurus site..." - cd ${{ runner.temp }}/llama-stack/docs + cd "${{ env.TEMP_DIR }}/llama-stack/docs" + + # Generate API docs for current build + npm run gen-api-docs all + + # Build the site npm run build + echo "โœ… Docusaurus build completed" - - name: Deploy to docs directory + - name: Deploy to GitHub Pages run: | echo "๐Ÿ—‚๏ธ Deploying Docusaurus build..." - # Smart deployment: clear everything except legacy, .git, and .nojekyll - find ${{ github.workspace }}/docs -mindepth 1 -maxdepth 1 ! -name 'legacy' ! -name '.git' ! -name '.nojekyll' -exec rm -rf {} + + DOCS_DIR="${{ github.workspace }}/docs" + + # Smart deployment: clear everything except .git, .nojekyll, and archived versions + find "$DOCS_DIR" -mindepth 1 -maxdepth 1 ! -name '.git' ! -name '.nojekyll' ! -name 'v[0-9]*' -exec rm -rf {} + # Copy Docusaurus build output - cp -r ${{ runner.temp }}/llama-stack/docs/build/* ${{ github.workspace }}/docs/ + cp -r "${{ env.TEMP_DIR }}/llama-stack/docs/build/"* "$DOCS_DIR/" # Ensure .nojekyll exists - touch "${{ github.workspace }}/docs/.nojekyll" + touch "$DOCS_DIR/.nojekyll" echo "โœ… Docusaurus content deployed" - - name: Update version management + - name: Update workspace artifacts + if: env.BUILDING_LATEST != 'true' run: | - echo "โš™๏ธ Updating version management..." + echo "๐Ÿ“‹ Updating workspace versioning artifacts..." - # Copy/update versioning files to deployment - cp "${{ runner.temp }}/llama-stack/docs/versionsArchived.json" "${{ github.workspace }}/docs/" + BUILD_PATH="${{ env.TEMP_DIR }}/llama-stack/docs" - # Update versions.json if we built a new version - if [ "${{ env.BUILDING_LATEST }}" != "true" ]; then - echo "Adding version ${{ env.VERSION_TAG }} to versions.json" + # Copy versioned docs and sidebars to workspace (for git commit) + if [ -d "$BUILD_PATH/versioned_docs" ]; then + cp -r "$BUILD_PATH/versioned_docs" "${{ github.workspace }}/" + echo "โœ… Copied versioned_docs to workspace" + fi - # Load current versions.json - VERSIONS_FILE="${{ github.workspace }}/docs/versions.json" - if [ -f "$VERSIONS_FILE" ]; then - CURRENT_VERSIONS=$(cat "$VERSIONS_FILE") - else - CURRENT_VERSIONS="[]" - fi + if [ -d "$BUILD_PATH/versioned_sidebars" ]; then + cp -r "$BUILD_PATH/versioned_sidebars" "${{ github.workspace }}/" + echo "โœ… Copied versioned_sidebars to workspace" + fi - # Add new version if not already present - NEW_VERSION="${{ env.VERSION_TAG }}" - UPDATED_VERSIONS=$(echo "$CURRENT_VERSIONS" | jq --arg version "$NEW_VERSION" '. as $arr | if ($arr | index($version)) then $arr else [$version] + $arr end') + # Copy updated versions.json to workspace + cp "$BUILD_PATH/versions.json" "${{ github.workspace }}/" + echo "โœ… Updated workspace versions.json" - echo "$UPDATED_VERSIONS" > "$VERSIONS_FILE" - echo "โœ… Updated versions.json with $NEW_VERSION" - fi + - name: Setup versioning files in deployed site + run: | + echo "โš™๏ธ Setting up versioning configuration files..." - cp "${{ runner.temp }}/llama-stack/docs/versions.json" "${{ github.workspace }}/docs/" 2>/dev/null || echo "[]" > "${{ github.workspace }}/docs/versions.json" + BUILD_PATH="${{ env.TEMP_DIR }}/llama-stack/docs" - echo "โœ… Version management updated" + # Copy versioning files to deployment + cp "$BUILD_PATH/versionsArchived.json" "${{ github.workspace }}/docs/" + cp "$BUILD_PATH/versions.json" "${{ github.workspace }}/docs/" + + echo "โœ… Versioning files created" - name: Verify deployment structure run: | echo "๐Ÿ” Verifying deployment structure..." echo "Contents of docs directory:" - ls -la ${{ github.workspace }}/docs/ | head -10 - - echo -e "\nLegacy versions:" - ls -la ${{ github.workspace }}/docs/legacy/ 2>/dev/null | head -5 || echo "โŒ Legacy directory missing" + ls -la "${{ github.workspace }}/docs/" | head -10 echo -e "\nVersioning files:" [ -f "${{ github.workspace }}/docs/versionsArchived.json" ] && echo "โœ… versionsArchived.json exists" || echo "โŒ versionsArchived.json missing" [ -f "${{ github.workspace }}/docs/versions.json" ] && echo "โœ… versions.json exists" || echo "โŒ versions.json missing" - echo -e "\nDocusaurus versions:" - if [ -d "${{ github.workspace }}/docs/docs" ]; then - ls -la "${{ github.workspace }}/docs/docs/" | head -5 - fi - echo -e "\nโœ… Structure verification complete" - - name: Commit and push changes + - name: Commit versioning artifacts + if: env.BUILDING_LATEST != 'true' run: | - echo "๐Ÿ’พ Committing changes..." + echo "๐Ÿ’พ Committing versioning artifacts..." - cd ${{ github.workspace }} + cd "${{ github.workspace }}" git config --local user.email "github-actions[bot]@users.noreply.github.com" git config --local user.name "github-actions[bot]" - # Add all changes - git add . - - # Only commit if there are changes - if ! git diff --staged --quiet; then - if [ "${{ env.BUILDING_LATEST }}" = "true" ]; then - git commit -m "Update Docusaurus documentation (latest) + # Add versioning artifacts + git add versioned_docs/ versioned_sidebars/ versions.json 2>/dev/null || true - - Updated latest Docusaurus build - - Applied configuration patches for version dropdown - - Maintained legacy versions in /legacy/ directory - - Version dropdown shows 'Latest' with archived versions + # Add deployment + git add docs/ - [skip ci]" - else - git commit -m "Add Docusaurus version ${{ env.VERSION_TAG }} + # Commit if there are changes + if ! git diff --staged --quiet; then + git commit -m "Add Docusaurus version ${{ env.VERSION_TAG }} - - Created new Docusaurus version ${{ env.VERSION_TAG }} + - Created version snapshot in versioned_docs/version-${{ env.VERSION_TAG }}/ - Updated versions.json with new version - - Applied configuration patches for version dropdown - - Maintained legacy versions in /legacy/ directory + - Built and deployed multi-version site + + ๐Ÿค– Generated by Docusaurus versioning workflow" - [skip ci]" - fi git push echo "โœ… Changes committed and pushed" else @@ -397,7 +341,7 @@ jobs: with: path: 'docs' - deploy-after-build: + deploy: if: ${{ github.event.inputs.action == 'build-and-deploy' }} runs-on: ubuntu-latest needs: build diff --git a/.gitignore b/.gitignore index 4e77cb2..e238729 100644 --- a/.gitignore +++ b/.gitignore @@ -1,29 +1,7 @@ -# ============================================================================ -# Docusaurus Migration - Updated .gitignore -# ============================================================================ - -# PRESERVE: Legacy Sphinx versions (frozen documentation) -# โœ… docs/legacy/v0.2.11/ through docs/legacy/v0.2.22/ - TRACKED -# โœ… docs/versionsArchived.json - TRACKED -# โœ… docs/index.html - TRACKED (root redirect) -# โœ… docs/.nojekyll - TRACKED (GitHub Pages config) - -# IGNORE: Generated Docusaurus build output (rebuilt from main repo) -docs/docs/ -docs/assets/ -docs/js/ -docs/css/ -docs/img/ -docs/fonts/ -docs/*.js -docs/*.css -docs/*.map -docs/manifest.json -docs/sitemap.xml -docs/robots.txt -docs/404.html +# Generated Docusaurus build output +docs/ -# IGNORE: Node.js dependencies and build artifacts +# Node.js dependencies and build artifacts node_modules/ npm-debug.log* yarn-debug.log* @@ -33,15 +11,15 @@ yarn-error.log* package-lock.json yarn.lock -# IGNORE: Docusaurus build cache +# Docusaurus build cache .docusaurus/ build/ -# IGNORE: Legacy Sphinx build artifacts (keep content, ignore build files) +# Legacy Sphinx build artifacts _build/ .sphinx/ -# IGNORE: Python artifacts +# Python artifacts __pycache__/ *.py[cod] *$py.class @@ -51,37 +29,23 @@ env/ venv/ .venv/ -# IGNORE: IDE files +# IDE files .vscode/ .idea/ *.swp *.swo *~ -# IGNORE: OS files +# OS files .DS_Store Thumbs.db -# IGNORE: Logs and temporary files +# Logs and temporary files *.log # Docusaurus build artefacts docs/ -# Docusaurus build cache -.docusaurus/ -build/ - -# Node.js dependencies and build artifacts -node_modules/ -npm-debug.log* -yarn-debug.log* -yarn-error.log* -.npm -.yarn-integrity -package-lock.json -yarn.lock - # Claude CLAUDE.md .claude/settings.local.json diff --git a/local-build-test.sh b/local-build-test.sh deleted file mode 100755 index f7888bc..0000000 --- a/local-build-test.sh +++ /dev/null @@ -1,228 +0,0 @@ -#!/bin/bash -set -e - -# Local Build and Test Script for Llama Stack Documentation -# Simplified version - legacy docs are already in place - -echo "๐Ÿš€ Starting local Llama Stack documentation build..." - -# Configuration -TEMP_DIR=$(mktemp -d) -REPO_URL="https://github.com/llamastack/llama-stack.git" -DOCS_DIR="$(pwd)/docs" - -cleanup() { - echo "๐Ÿงน Cleaning up temporary directory..." - rm -rf "$TEMP_DIR" -} -trap cleanup EXIT - -# Step 1: Clone llama-stack repository -echo "๐Ÿ“ฅ Cloning llama-stack repository..." -git clone "$REPO_URL" "$TEMP_DIR/llama-stack" -echo "โœ… Repository cloned" - -# Step 2: Install dependencies -echo "๐Ÿ“ฆ Installing Docusaurus dependencies..." -cd "$TEMP_DIR/llama-stack/docs" -npm ci -echo "โœ… Dependencies installed" - -# Step 3: Apply configuration patches -echo "โš™๏ธ Applying Docusaurus configuration patches..." - -# Create versioning files -cat > "$TEMP_DIR/llama-stack/docs/versionsArchived.json" << 'EOF' -{ - "v0.2.22": "https://llamastack.github.io/legacy/v0.2.22/", - "v0.2.21": "https://llamastack.github.io/legacy/v0.2.21/", - "v0.2.20": "https://llamastack.github.io/legacy/v0.2.20/", - "v0.2.19": "https://llamastack.github.io/legacy/v0.2.19/", - "v0.2.18": "https://llamastack.github.io/legacy/v0.2.18/", - "v0.2.17": "https://llamastack.github.io/legacy/v0.2.17/", - "v0.2.16": "https://llamastack.github.io/legacy/v0.2.16/", - "v0.2.15": "https://llamastack.github.io/legacy/v0.2.15/", - "v0.2.14": "https://llamastack.github.io/legacy/v0.2.14/", - "v0.2.13": "https://llamastack.github.io/legacy/v0.2.13/", - "v0.2.12": "https://llamastack.github.io/legacy/v0.2.12/", - "v0.2.11": "https://llamastack.github.io/legacy/v0.2.11/" -} -EOF - -cat > "$TEMP_DIR/llama-stack/docs/versions.json" << 'EOF' -[] -EOF - -# Patch docusaurus.config.ts -echo "๐Ÿ”ง Patching Docusaurus configuration..." - -# Apply comprehensive patches to docusaurus.config.ts -cat > "$TEMP_DIR/config-patch.js" << 'EOF' -const fs = require('fs'); -const path = require('path'); - -const configPath = process.argv[2]; -let config = fs.readFileSync(configPath, 'utf8'); - -// Add archived versions loading at the top -const versioningImports = ` -// Import archived versions configuration -const fs = require('fs'); -const path = require('path'); - -// Load archived versions -const versionsArchived = (() => { - try { - return JSON.parse(fs.readFileSync(path.join(__dirname, 'versionsArchived.json'), 'utf8')); - } catch (e) { - console.warn('Could not load versionsArchived.json:', e); - return {}; - } -})(); - -// Create dropdown items for archived versions -const archivedVersionsDropdownItems = Object.entries(versionsArchived).map( - ([versionName, versionUrl]) => ({ - label: versionName, - href: versionUrl, - }) -); -`; - -// Insert versioning imports after existing imports -config = config.replace( - /import type \* as OpenApiPlugin from "docusaurus-plugin-openapi-docs";/, - `import type * as OpenApiPlugin from "docusaurus-plugin-openapi-docs"; - -${versioningImports}` -); - -// Change version label from "Next ๐Ÿšง" to "Latest" -config = config.replace( - /label: 'Next ๐Ÿšง'/, - "label: 'Latest'" -); - -// Add versioning configuration to docs config -config = config.replace( - /docItemComponent: "@theme\/ApiItem", \/\/ Derived from docusaurus-theme-openapi/, - `docItemComponent: "@theme/ApiItem", // Derived from docusaurus-theme-openapi - - // Versioning configuration - lastVersion: 'current', - versions: { - current: { - label: 'Latest', - path: '', - }, - }, - - // Only include current version since we handle archived versions separately - onlyIncludeVersions: ['current'],` -); - -// Add version dropdown to navbar items (position it on the right before GitHub) -const versionDropdown = ` { - href: 'https://github.com/llamastack/llama-stack', - label: 'GitHub', - position: 'right', - }, - // Version dropdown with archived versions - { - type: 'docsVersionDropdown', - position: 'right', - dropdownItemsAfter: archivedVersionsDropdownItems.length > 0 ? [ - { - type: 'html', - value: '', - }, - { - type: 'html', - className: 'dropdown-archived-versions', - value: 'Archived versions', - }, - ...archivedVersionsDropdownItems, - { - type: 'html', - value: '', - }, - { - to: '/versions', - label: 'All versions', - }, - ] : [], - },`; - -// Replace GitHub item with version dropdown + GitHub -config = config.replace( - / {\s*href: 'https:\/\/github\.com\/llamastack\/llama-stack',\s*label: 'GitHub',\s*position: 'right',\s*},/, - versionDropdown -); - -fs.writeFileSync(configPath, config); -console.log('โœ… Docusaurus configuration patched successfully'); -EOF - -node "$TEMP_DIR/config-patch.js" "$TEMP_DIR/llama-stack/docs/docusaurus.config.ts" - -echo "โœ… Configuration patches applied" - -# Step 4: Generate API documentation -echo "๐Ÿ“š Generating API documentation..." -npm run gen-api-docs all -echo "โœ… API docs generated" - -# Step 5: Build Docusaurus site -echo "๐Ÿ—๏ธ Building Docusaurus site..." -npm run build -echo "โœ… Docusaurus build completed" - -# Step 6: Deploy to docs directory (preserve legacy and .git) -echo "๐Ÿ—‚๏ธ Deploying Docusaurus build..." - -# Smart deployment: clear everything except legacy, .git, and .nojekyll -find "$DOCS_DIR" -mindepth 1 -maxdepth 1 ! -name 'legacy' ! -name '.git' ! -name '.nojekyll' -exec rm -rf {} + - -# Copy Docusaurus build output -cp -r "$TEMP_DIR/llama-stack/docs/build/"* "$DOCS_DIR/" - -# Ensure .nojekyll exists (in case it didn't exist before) -touch "$DOCS_DIR/.nojekyll" - -echo "โœ… Docusaurus content deployed" - -# Step 7: Create versioning configuration files in deployed site -echo "โš™๏ธ Setting up versioning configuration files..." - -# Copy versioning files to deployment -cp "$TEMP_DIR/llama-stack/docs/versionsArchived.json" "$DOCS_DIR/" -cp "$TEMP_DIR/llama-stack/docs/versions.json" "$DOCS_DIR/" - -echo "โœ… Versioning files created" - -# Step 8: Verify deployment structure -echo "๐Ÿ” Verifying deployment structure..." - -echo "Contents of docs directory:" -ls -la "$DOCS_DIR/" | head -10 - -echo -e "\nLegacy versions:" -ls -la "$DOCS_DIR/legacy/" 2>/dev/null | head -5 || echo "โŒ Legacy directory missing" - -echo -e "\nVersioning files:" -[ -f "$DOCS_DIR/versionsArchived.json" ] && echo "โœ… versionsArchived.json exists" || echo "โŒ versionsArchived.json missing" -[ -f "$DOCS_DIR/versions.json" ] && echo "โœ… versions.json exists" || echo "โŒ versions.json missing" - -echo -e "\nโœ… Structure verification complete" - -# Step 9: Start local server for testing -echo "๐ŸŒ Starting local development server..." -echo "๐Ÿ“ Your documentation is available at: http://localhost:3000" -echo "๐Ÿ”— Main docs: http://localhost:3000/docs.html" -echo "๐Ÿ“š API Reference: http://localhost:3000/docs/api/llama-stack-specification" -echo "๐Ÿ“š Legacy versions: http://localhost:3000/legacy/" -echo "" -echo "Press Ctrl+C to stop the server" - -cd "$DOCS_DIR" -python3 -m http.server 3000 2>/dev/null || python -m SimpleHTTPServer 3000 diff --git a/run-workflow.py b/run-workflow.py index 7e08473..713476c 100755 --- a/run-workflow.py +++ b/run-workflow.py @@ -42,11 +42,18 @@ def load_workflow(self) -> Dict[str, Any]: def setup_runner_context(self): """Set up GitHub Actions runner context variables""" self.temp_dir = tempfile.mkdtemp(prefix="workflow-runner-") + self.github_env_file = os.path.join(self.temp_dir, "github_env") + + # Create empty GITHUB_ENV file + with open(self.github_env_file, "w") as _: + pass + self.env.update( { "RUNNER_TEMP": self.temp_dir, "GITHUB_WORKSPACE": os.getcwd(), "CI": "true", + "GITHUB_ENV": self.github_env_file, } ) print(f"๐Ÿ“ Temp directory: {self.temp_dir}") @@ -58,6 +65,19 @@ def cleanup(self): def replace_variables(self, text: str) -> str: """Replace GitHub Actions variables with their values""" + # Load any environment variables from GITHUB_ENV file + github_env_vars = {} + if hasattr(self, "github_env_file") and os.path.exists(self.github_env_file): + try: + with open(self.github_env_file, "r") as f: + for line in f: + line = line.strip() + if "=" in line: + key, value = line.split("=", 1) + github_env_vars[key] = value + except Exception: + pass + replacements = { "${{ runner.temp }}": self.env.get("RUNNER_TEMP", ""), "${{ github.workspace }}": self.env.get("GITHUB_WORKSPACE", ""), @@ -67,7 +87,11 @@ def replace_variables(self, text: str) -> str: "${{ github.event.inputs.version || 'latest' }}": self.env.get( "INPUT_VERSION", "latest" ), + "${{ env.TEMP_DIR }}": github_env_vars.get("TEMP_DIR", ""), + "${{ env.BUILDING_LATEST }}": github_env_vars.get("BUILDING_LATEST", ""), + "${{ env.VERSION_TAG }}": github_env_vars.get("VERSION_TAG", ""), } + for key, value in replacements.items(): text = text.replace(key, value) return text @@ -135,6 +159,17 @@ def run_command(self, cmd: str, name: str = None) -> bool: if not cmd.startswith("source .venv/bin/activate"): cmd = f"source .venv/bin/activate && {cmd}" + # Source GITHUB_ENV file to make environment variables available + if hasattr(self, "github_env_file") and os.path.exists(self.github_env_file): + # Check if file has content + try: + with open(self.github_env_file, "r") as f: + content = f.read().strip() + if content: + cmd = f"source {self.github_env_file} && {cmd}" + except Exception: + pass + try: result = subprocess.run( cmd, @@ -167,9 +202,15 @@ def should_skip_step(self, step: Dict[str, Any]) -> bool: if pattern in name.lower(): return True - # Skip steps that use GitHub Actions - if "uses" in step and "setup-uv" not in step.get("uses", ""): - return True + # Skip steps that use GitHub Actions (except setup actions we can handle) + if "uses" in step: + uses = step.get("uses", "") + # Allow setup actions we can handle + if not any( + allowed in uses + for allowed in ["setup-uv", "setup-node", "actions/setup-node"] + ): + return True return False @@ -200,6 +241,20 @@ def run_step(self, step: Dict[str, Any]) -> bool: "curl -LsSf https://astral.sh/uv/install.sh | sh", None ) + # Handle setup-node action + if "uses" in step and ( + "setup-node" in step["uses"] or "actions/setup-node" in step["uses"] + ): + print(f"\n๐Ÿ”ง {name}") + # Check if node is installed + if shutil.which("node") and shutil.which("npm"): + print(" โœ“ Node.js and npm are already installed") + return True + else: + print(" โš ๏ธ Node.js/npm not found. Please install Node.js 20+ first.") + print(" Visit: https://nodejs.org/") + return False + return True def run(self): @@ -270,15 +325,20 @@ def main(): args = parser.parse_args() # Check for required tools - if not shutil.which("uv"): - print("โŒ uv is not installed. Please install it first:") - print(" curl -LsSf https://astral.sh/uv/install.sh | sh") - sys.exit(1) - if not shutil.which("git"): print("โŒ git is not installed. Please install git first.") sys.exit(1) + if not shutil.which("node"): + print("โŒ Node.js is not installed. Please install Node.js 20+ first:") + print(" Visit: https://nodejs.org/") + sys.exit(1) + + if not shutil.which("npm"): + print("โŒ npm is not installed. Please install Node.js (includes npm) first:") + print(" Visit: https://nodejs.org/") + sys.exit(1) + # Check if workflow file exists workflow_file = ".github/workflows/build-and-deploy.yml" if not os.path.exists(workflow_file): diff --git a/versioned_docs/resources/prompt-format.png b/versioned_docs/resources/prompt-format.png new file mode 100644 index 0000000..afcd076 Binary files /dev/null and b/versioned_docs/resources/prompt-format.png differ diff --git a/versioned_docs/version-v0.2.23/advanced_apis/evaluation.mdx b/versioned_docs/version-v0.2.23/advanced_apis/evaluation.mdx new file mode 100644 index 0000000..1efaa4c --- /dev/null +++ b/versioned_docs/version-v0.2.23/advanced_apis/evaluation.mdx @@ -0,0 +1,163 @@ +# Evaluation + +## Evaluation Concepts + +The Llama Stack Evaluation flow allows you to run evaluations on your GenAI application datasets or pre-registered benchmarks. + +We introduce a set of APIs in Llama Stack for supporting running evaluations of LLM applications: +- `/datasetio` + `/datasets` API +- `/scoring` + `/scoring_functions` API +- `/eval` + `/benchmarks` API + +This guide goes over the sets of APIs and developer experience flow of using Llama Stack to run evaluations for different use cases. Checkout our Colab notebook on working examples with evaluations [here](https://colab.research.google.com/drive/10CHyykee9j2OigaIcRv47BKG9mrNm0tJ?usp=sharing). + +The Evaluation APIs are associated with a set of Resources. Please visit the Resources section in our [Core Concepts](../concepts/index.mdx) guide for better high-level understanding. + +- **DatasetIO**: defines interface with datasets and data loaders. + - Associated with `Dataset` resource. +- **Scoring**: evaluate outputs of the system. + - Associated with `ScoringFunction` resource. We provide a suite of out-of-the box scoring functions and also the ability for you to add custom evaluators. These scoring functions are the core part of defining an evaluation task to output evaluation metrics. +- **Eval**: generate outputs (via Inference or Agents) and perform scoring. + - Associated with `Benchmark` resource. + +## Evaluation Providers + +Llama Stack provides multiple evaluation providers: + +- **Meta Reference** (`inline::meta-reference`) - Meta's reference implementation with multi-language support +- **NVIDIA** (`remote::nvidia`) - NVIDIA's evaluation platform integration + +### Meta Reference + +Meta's reference implementation of evaluation tasks with support for multiple languages and evaluation metrics. + +#### Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `RedisKVStoreConfig \| SqliteKVStoreConfig \| PostgresKVStoreConfig \| MongoDBKVStoreConfig` | No | sqlite | Key-value store configuration | + +#### Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/meta_reference_eval.db +``` + +#### Features + +- Multi-language evaluation support +- Comprehensive evaluation metrics +- Integration with various key-value stores (SQLite, Redis, PostgreSQL, MongoDB) +- Built-in support for popular benchmarks + +### NVIDIA + +NVIDIA's evaluation provider for running evaluation tasks on NVIDIA's platform. + +#### Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `evaluator_url` | `str` | No | http://0.0.0.0:7331 | The url for accessing the evaluator service | + +#### Sample Configuration + +```yaml +evaluator_url: ${env.NVIDIA_EVALUATOR_URL:=http://localhost:7331} +``` + +#### Features + +- Integration with NVIDIA's evaluation platform +- Remote evaluation capabilities +- Scalable evaluation processing + +## Open-benchmark Eval + +### List of open-benchmarks Llama Stack support + +Llama stack pre-registers several popular open-benchmarks to easily evaluate model performance via CLI. + +The list of open-benchmarks we currently support: +- [MMLU-COT](https://arxiv.org/abs/2009.03300) (Measuring Massive Multitask Language Understanding): Benchmark designed to comprehensively evaluate the breadth and depth of a model's academic and professional understanding +- [GPQA-COT](https://arxiv.org/abs/2311.12022) (A Graduate-Level Google-Proof Q&A Benchmark): A challenging benchmark of 448 multiple-choice questions written by domain experts in biology, physics, and chemistry. +- [SimpleQA](https://openai.com/index/introducing-simpleqa/): Benchmark designed to access models to answer short, fact-seeking questions. +- [MMMU](https://arxiv.org/abs/2311.16502) (A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI): Benchmark designed to evaluate multimodal models. + +You can follow this [contributing guide](../references/evals_reference/index.mdx#open-benchmark-contributing-guide) to add more open-benchmarks to Llama Stack + +### Run evaluation on open-benchmarks via CLI + +We have built-in functionality to run the supported open-benchmarks using llama-stack-client CLI + +#### Spin up Llama Stack server + +Spin up llama stack server with 'open-benchmark' template +``` +llama stack run llama_stack/distributions/open-benchmark/run.yaml + +``` + +#### Run eval CLI +There are 3 necessary inputs to run a benchmark eval +- `list of benchmark_ids`: The list of benchmark ids to run evaluation on +- `model-id`: The model id to evaluate on +- `output_dir`: Path to store the evaluate results +``` +llama-stack-client eval run-benchmark ... \ +--model_id \ +--output_dir +``` + +You can run +``` +llama-stack-client eval run-benchmark help +``` +to see the description of all the flags that eval run-benchmark has + +In the output log, you can find the file path that has your evaluation results. Open that file and you can see you aggregate evaluation results over there. + +## Usage Example + +Here's a basic example of using the evaluation API: + +```python +from llama_stack_client import LlamaStackClient + +client = LlamaStackClient(base_url="http://localhost:8321") + +# Register a dataset for evaluation +client.datasets.register( + purpose="evaluation", + source={ + "type": "uri", + "uri": "huggingface://datasets/llamastack/evaluation_dataset" + }, + dataset_id="my_eval_dataset" +) + +# Run evaluation +eval_result = client.eval.run_evaluation( + dataset_id="my_eval_dataset", + scoring_functions=["accuracy", "bleu"], + model_id="my_model" +) + +print(f"Evaluation completed: {eval_result}") +``` + +## Best Practices + +- **Choose appropriate providers**: Use Meta Reference for comprehensive evaluation, NVIDIA for platform-specific needs +- **Configure storage properly**: Ensure your key-value store configuration matches your performance requirements +- **Monitor evaluation progress**: Large evaluations can take time - implement proper monitoring +- **Use appropriate scoring functions**: Select scoring metrics that align with your evaluation goals + +## What's Next? + +- Check out our Colab notebook on working examples with running benchmark evaluations [here](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb#scrollTo=mxLCsP4MvFqP). +- Check out our [Building Applications - Evaluation](../building_applications/evals.mdx) guide for more details on how to use the Evaluation APIs to evaluate your applications. +- Check out our [Evaluation Reference](../references/evals_reference/index.mdx) for more details on the APIs. +- Explore the [Scoring](./scoring.mdx) documentation for available scoring functions. diff --git a/versioned_docs/version-v0.2.23/advanced_apis/post_training.mdx b/versioned_docs/version-v0.2.23/advanced_apis/post_training.mdx new file mode 100644 index 0000000..516ac07 --- /dev/null +++ b/versioned_docs/version-v0.2.23/advanced_apis/post_training.mdx @@ -0,0 +1,305 @@ +# Post-Training + +Post-training in Llama Stack allows you to fine-tune models using various providers and frameworks. This section covers all available post-training providers and how to use them effectively. + +## Overview + +Llama Stack provides multiple post-training providers: + +- **HuggingFace SFTTrainer** (`inline::huggingface`) - Fine-tuning using HuggingFace ecosystem +- **TorchTune** (`inline::torchtune`) - Fine-tuning using Meta's TorchTune framework +- **NVIDIA** (`remote::nvidia`) - Fine-tuning using NVIDIA's platform + +## HuggingFace SFTTrainer + +[HuggingFace SFTTrainer](https://huggingface.co/docs/trl/en/sft_trainer) is an inline post training provider for Llama Stack. It allows you to run supervised fine tuning on a variety of models using many datasets. + +### Features + +- Simple access through the post_training API +- Fully integrated with Llama Stack +- GPU support, CPU support, and MPS support (MacOS Metal Performance Shaders) + +### Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `device` | `str` | No | cuda | | +| `distributed_backend` | `Literal['fsdp', 'deepspeed']` | No | | | +| `checkpoint_format` | `Literal['full_state', 'huggingface']` | No | huggingface | | +| `chat_template` | `str` | No | | +| `model_specific_config` | `dict` | No | `{'trust_remote_code': True, 'attn_implementation': 'sdpa'}` | | +| `max_seq_length` | `int` | No | 2048 | | +| `gradient_checkpointing` | `bool` | No | False | | +| `save_total_limit` | `int` | No | 3 | | +| `logging_steps` | `int` | No | 10 | | +| `warmup_ratio` | `float` | No | 0.1 | | +| `weight_decay` | `float` | No | 0.01 | | +| `dataloader_num_workers` | `int` | No | 4 | | +| `dataloader_pin_memory` | `bool` | No | True | | + +### Sample Configuration + +```yaml +checkpoint_format: huggingface +distributed_backend: null +device: cpu +``` + +### Setup + +You can access the HuggingFace trainer via the `starter` distribution: + +```bash +llama stack build --distro starter --image-type venv +llama stack run --image-type venv ~/.llama/distributions/starter/starter-run.yaml +``` + +### Usage Example + +```python +import time +import uuid + +from llama_stack_client.types import ( + post_training_supervised_fine_tune_params, + algorithm_config_param, +) + +def create_http_client(): + from llama_stack_client import LlamaStackClient + return LlamaStackClient(base_url="http://localhost:8321") + +client = create_http_client() + +# Example Dataset +client.datasets.register( + purpose="post-training/messages", + source={ + "type": "uri", + "uri": "huggingface://datasets/llamastack/simpleqa?split=train", + }, + dataset_id="simpleqa", +) + +training_config = post_training_supervised_fine_tune_params.TrainingConfig( + data_config=post_training_supervised_fine_tune_params.TrainingConfigDataConfig( + batch_size=32, + data_format="instruct", + dataset_id="simpleqa", + shuffle=True, + ), + gradient_accumulation_steps=1, + max_steps_per_epoch=0, + max_validation_steps=1, + n_epochs=4, +) + +algorithm_config = algorithm_config_param.LoraFinetuningConfig( + alpha=1, + apply_lora_to_mlp=True, + apply_lora_to_output=False, + lora_attn_modules=["q_proj"], + rank=1, + type="LoRA", +) + +job_uuid = f"test-job{uuid.uuid4()}" + +# Example Model +training_model = "ibm-granite/granite-3.3-8b-instruct" + +start_time = time.time() +response = client.post_training.supervised_fine_tune( + job_uuid=job_uuid, + logger_config={}, + model=training_model, + hyperparam_search_config={}, + training_config=training_config, + algorithm_config=algorithm_config, + checkpoint_dir="output", +) +print("Job: ", job_uuid) + +# Wait for the job to complete! +while True: + status = client.post_training.job.status(job_uuid=job_uuid) + if not status: + print("Job not found") + break + + print(status) + if status.status == "completed": + break + + print("Waiting for job to complete...") + time.sleep(5) + +end_time = time.time() +print("Job completed in", end_time - start_time, "seconds!") + +print("Artifacts:") +print(client.post_training.job.artifacts(job_uuid=job_uuid)) +``` + +## TorchTune + +[TorchTune](https://github.com/pytorch/torchtune) is an inline post training provider for Llama Stack. It provides a simple and efficient way to fine-tune language models using PyTorch. + +### Features + +- Simple access through the post_training API +- Fully integrated with Llama Stack +- GPU support and single device capabilities +- Support for LoRA + +### Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `torch_seed` | `int \| None` | No | | | +| `checkpoint_format` | `Literal['meta', 'huggingface']` | No | meta | | + +### Sample Configuration + +```yaml +checkpoint_format: meta +``` + +### Setup + +You can access the TorchTune trainer by writing your own yaml pointing to the provider: + +```yaml +post_training: + - provider_id: torchtune + provider_type: inline::torchtune + config: {} +``` + +You can then build and run your own stack with this provider. + +### Usage Example + +```python +import time +import uuid + +from llama_stack_client.types import ( + post_training_supervised_fine_tune_params, + algorithm_config_param, +) + +def create_http_client(): + from llama_stack_client import LlamaStackClient + return LlamaStackClient(base_url="http://localhost:8321") + +client = create_http_client() + +# Example Dataset +client.datasets.register( + purpose="post-training/messages", + source={ + "type": "uri", + "uri": "huggingface://datasets/llamastack/simpleqa?split=train", + }, + dataset_id="simpleqa", +) + +training_config = post_training_supervised_fine_tune_params.TrainingConfig( + data_config=post_training_supervised_fine_tune_params.TrainingConfigDataConfig( + batch_size=32, + data_format="instruct", + dataset_id="simpleqa", + shuffle=True, + ), + gradient_accumulation_steps=1, + max_steps_per_epoch=0, + max_validation_steps=1, + n_epochs=4, +) + +algorithm_config = algorithm_config_param.LoraFinetuningConfig( + alpha=1, + apply_lora_to_mlp=True, + apply_lora_to_output=False, + lora_attn_modules=["q_proj"], + rank=1, + type="LoRA", +) + +job_uuid = f"test-job{uuid.uuid4()}" + +# Example Model +training_model = "meta-llama/Llama-2-7b-hf" + +start_time = time.time() +response = client.post_training.supervised_fine_tune( + job_uuid=job_uuid, + logger_config={}, + model=training_model, + hyperparam_search_config={}, + training_config=training_config, + algorithm_config=algorithm_config, + checkpoint_dir="output", +) +print("Job: ", job_uuid) + +# Wait for the job to complete! +while True: + status = client.post_training.job.status(job_uuid=job_uuid) + if not status: + print("Job not found") + break + + print(status) + if status.status == "completed": + break + + print("Waiting for job to complete...") + time.sleep(5) + +end_time = time.time() +print("Job completed in", end_time - start_time, "seconds!") + +print("Artifacts:") +print(client.post_training.job.artifacts(job_uuid=job_uuid)) +``` + +## NVIDIA + +NVIDIA's post-training provider for fine-tuning models on NVIDIA's platform. + +### Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The NVIDIA API key. | +| `dataset_namespace` | `str \| None` | No | default | The NVIDIA dataset namespace. | +| `project_id` | `str \| None` | No | test-example-model@v1 | The NVIDIA project ID. | +| `customizer_url` | `str \| None` | No | | Base URL for the NeMo Customizer API | +| `timeout` | `int` | No | 300 | Timeout for the NVIDIA Post Training API | +| `max_retries` | `int` | No | 3 | Maximum number of retries for the NVIDIA Post Training API | +| `output_model_dir` | `str` | No | test-example-model@v1 | Directory to save the output model | + +### Sample Configuration + +```yaml +api_key: ${env.NVIDIA_API_KEY:=} +dataset_namespace: ${env.NVIDIA_DATASET_NAMESPACE:=default} +project_id: ${env.NVIDIA_PROJECT_ID:=test-project} +customizer_url: ${env.NVIDIA_CUSTOMIZER_URL:=http://nemo.test} +``` + +## Best Practices + +- **Choose the right provider**: Use HuggingFace for broader compatibility, TorchTune for Meta models, or NVIDIA for their ecosystem +- **Configure hardware appropriately**: Ensure your configuration matches your available hardware (CPU, GPU, MPS) +- **Monitor jobs**: Always monitor job status and handle completion appropriately +- **Use appropriate datasets**: Ensure your dataset format matches the expected input format for your chosen provider + +## Next Steps + +- Check out the [Building Applications - Fine-tuning](../building_applications/index.mdx) guide for application-level examples +- See the [Providers](../providers/post_training/index.mdx) section for detailed provider documentation +- Review the [API Reference](../advanced_apis/post_training.mdx) for complete API documentation diff --git a/versioned_docs/version-v0.2.23/advanced_apis/scoring.mdx b/versioned_docs/version-v0.2.23/advanced_apis/scoring.mdx new file mode 100644 index 0000000..0ce787e --- /dev/null +++ b/versioned_docs/version-v0.2.23/advanced_apis/scoring.mdx @@ -0,0 +1,193 @@ +# Scoring + +The Scoring API in Llama Stack allows you to evaluate outputs of your GenAI system using various scoring functions and metrics. This section covers all available scoring providers and their configuration. + +## Overview + +Llama Stack provides multiple scoring providers: + +- **Basic** (`inline::basic`) - Simple evaluation metrics and scoring functions +- **Braintrust** (`inline::braintrust`) - Advanced evaluation using the Braintrust platform +- **LLM-as-Judge** (`inline::llm-as-judge`) - Uses language models to evaluate responses + +The Scoring API is associated with `ScoringFunction` resources and provides a suite of out-of-the-box scoring functions. You can also add custom evaluators to meet specific evaluation needs. + +## Basic Scoring + +Basic scoring provider for simple evaluation metrics and scoring functions. This provider offers fundamental scoring capabilities without external dependencies. + +### Configuration + +No configuration required - this provider works out of the box. + +```yaml +{} +``` + +### Features + +- Simple evaluation metrics (accuracy, precision, recall, F1-score) +- String matching and similarity metrics +- Basic statistical scoring functions +- No external dependencies required +- Fast execution for standard metrics + +### Use Cases + +- Quick evaluation of basic accuracy metrics +- String similarity comparisons +- Statistical analysis of model outputs +- Development and testing scenarios + +## Braintrust + +Braintrust scoring provider for evaluation and scoring using the [Braintrust platform](https://braintrustdata.com/). Braintrust provides advanced evaluation capabilities and experiment tracking. + +### Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `openai_api_key` | `str \| None` | No | | The OpenAI API Key for LLM-powered evaluations | + +### Sample Configuration + +```yaml +openai_api_key: ${env.OPENAI_API_KEY:=} +``` + +### Features + +- Advanced evaluation metrics +- Experiment tracking and comparison +- LLM-powered evaluation functions +- Integration with Braintrust's evaluation suite +- Detailed scoring analytics and insights + +### Use Cases + +- Production evaluation pipelines +- A/B testing of model versions +- Advanced scoring with custom metrics +- Detailed evaluation reporting and analysis + +## LLM-as-Judge + +LLM-as-judge scoring provider that uses language models to evaluate and score responses. This approach leverages the reasoning capabilities of large language models to assess quality, relevance, and other subjective metrics. + +### Configuration + +No configuration required - this provider works out of the box. + +```yaml +{} +``` + +### Features + +- Subjective quality evaluation using LLMs +- Flexible evaluation criteria definition +- Natural language evaluation explanations +- Support for complex evaluation scenarios +- Contextual understanding of responses + +### Use Cases + +- Evaluating response quality and relevance +- Assessing creativity and coherence +- Subjective metric evaluation +- Human-like judgment for complex tasks + +## Usage Examples + +### Basic Scoring Example + +```python +from llama_stack_client import LlamaStackClient + +client = LlamaStackClient(base_url="http://localhost:8321") + +# Register a basic accuracy scoring function +client.scoring_functions.register( + scoring_function_id="basic_accuracy", + provider_id="basic", + provider_scoring_function_id="accuracy" +) + +# Use the scoring function +result = client.scoring.score( + input_rows=[ + {"expected": "Paris", "actual": "Paris"}, + {"expected": "London", "actual": "Paris"} + ], + scoring_function_id="basic_accuracy" +) +print(f"Accuracy: {result.results[0].score}") +``` + +### LLM-as-Judge Example + +```python +# Register an LLM-as-judge scoring function +client.scoring_functions.register( + scoring_function_id="quality_judge", + provider_id="llm_judge", + provider_scoring_function_id="response_quality", + params={ + "criteria": "Evaluate response quality, relevance, and helpfulness", + "scale": "1-10" + } +) + +# Score responses using LLM judgment +result = client.scoring.score( + input_rows=[{ + "query": "What is machine learning?", + "response": "Machine learning is a subset of AI that enables computers to learn patterns from data..." + }], + scoring_function_id="quality_judge" +) +``` + +### Braintrust Integration Example + +```python +# Register a Braintrust scoring function +client.scoring_functions.register( + scoring_function_id="braintrust_eval", + provider_id="braintrust", + provider_scoring_function_id="semantic_similarity" +) + +# Run evaluation with Braintrust +result = client.scoring.score( + input_rows=[{ + "reference": "The capital of France is Paris", + "candidate": "Paris is the capital city of France" + }], + scoring_function_id="braintrust_eval" +) +``` + +## Best Practices + +- **Choose appropriate providers**: Use Basic for simple metrics, Braintrust for advanced analytics, LLM-as-Judge for subjective evaluation +- **Define clear criteria**: When using LLM-as-Judge, provide specific evaluation criteria and scales +- **Validate scoring functions**: Test your scoring functions with known examples before production use +- **Monitor performance**: Track scoring performance and adjust thresholds based on results +- **Combine multiple metrics**: Use different scoring providers together for comprehensive evaluation + +## Integration with Evaluation + +The Scoring API works closely with the [Evaluation](./evaluation.mdx) API to provide comprehensive evaluation workflows: + +1. **Datasets** are loaded via the DatasetIO API +2. **Evaluation** generates model outputs using the Eval API +3. **Scoring** evaluates the quality of outputs using various scoring functions +4. **Results** are aggregated and reported for analysis + +## Next Steps + +- Check out the [Evaluation](./evaluation.mdx) guide for running complete evaluations +- See the [Building Applications - Evaluation](../building_applications/evals.mdx) guide for application examples +- Review the [Evaluation Reference](../references/evals_reference/) for comprehensive scoring function usage +- Explore the [Evaluation Concepts](../concepts/evaluation_concepts) for detailed conceptual information diff --git a/versioned_docs/version-v0.2.23/api/agents.tag.mdx b/versioned_docs/version-v0.2.23/api/agents.tag.mdx new file mode 100644 index 0000000..ce0632f --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/agents.tag.mdx @@ -0,0 +1,24 @@ +--- +id: agents +title: "Agents API for creating and interacting with agentic systems." +description: "Agents API for creating and interacting with agentic systems." +custom_edit_url: null +--- + + + +Main functionalities provided by this API: +- Create agents with specific instructions and ability to use tools. +- Interactions with agents are grouped into sessions ("threads"), and each interaction is called a "turn". +- Agents can be provided with various tools (see the ToolGroups and ToolRuntime APIs for more details). +- Agents can be provided with various shields (see the Safety API for more details). +- Agents can also use Memory to retrieve information from knowledge bases. See the RAG Tool and Vector IO APIs for more details. + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/append-rows-to-a-dataset.api.mdx b/versioned_docs/version-v0.2.23/api/append-rows-to-a-dataset.api.mdx new file mode 100644 index 0000000..9622f68 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/append-rows-to-a-dataset.api.mdx @@ -0,0 +1,68 @@ +--- +id: append-rows-to-a-dataset +title: "Append rows to a dataset." +description: "Append rows to a dataset." +sidebar_label: "Append rows to a dataset." +hide_title: true +hide_table_of_contents: true +api: eJztV01v2zgQ/SuETi0Q22nQolvf0qaLDbaFA8c9LFJjMZLGFluK1JKUE0Pwf+8MKcVy7AQpsJcA9iGRqPme92akJrHoKqMdumTcJGenp/wvR5dZWXlpdDJOJn8nm5Pk7aFHswKFxf9qdF7cghNSr0DJXBgrSlALY0vMk5MkM9qj9qwPVaVkBqw/+uHYSJO4rMAS+MqvKySzJv2BmSfFypoKrZcxOufB164nJ8nsEi0J7ob112x2JaK0yEyOHL+XXmFP13kr9XJP9bO1FHwQPhEgXGGsF64uS7BrYRbCU8YYZG4LmRVCxqStBO0FJSxAt8+Dnw3b9yDVcz1HaXatjKbcRFGXoAcWIYdUoejp7ITDnqSmnHX2jCxfTcIVqNfiXHybfmmTySj4FEXtMBfeUGdJG1coSmOR0uR+hs4JSE3tg3NXYSYXMhMmy2prkdw/iIsCgzyX0d9Vr6ELUA5PEsaPtIST8U3X4a5b98Wb3/cvlumR4nVYFgtryhDD+dXlUExNvSzUmtqjlLl1YvrnJ/H+j9P3Qy4a3kFZRWR0+CKob/19hFxMI8S38TwT+ZsNZ//27MNh5mRKEitEQeqOL7wxpKzXnWG2SkBYyhXS/9LUJMO1lSUeSXUk1Qsj1dmHrb8ZIf0rI71llutT6x9TEyeoQniXIeZcNiYbeBRKltIPxZVCoIA8oQeWQCRR9NAOW769e2xTObQraj6Vk6mEVB4GVq3xjurt6Q7bMhypdaTWC6LWu/6+umRkU+TiOqJ967+l1/k+5Lti5EMxqQkJCGXYSinS4tHGU8HoWUuvHBdQK79PsacMH0l1JNXLItXpIW/PoFBgCUMNlo6zuQBPu8pfTjj+B4SpKtS5sBwOVRpEHmWHDHqwUCJRmY00iaYb0mgF/pVMKMk2KvDFXiF42V1edC1olYKL6DHs0+iVfW2L721NzdijYgugzWYehWljfzT5miX+J1ZzND0psBbWnKLH0h3SPgymJjEaJ4tQsVZD10pRh+/vU2Po5UH3j3Rdpmj7J12+Jw8C6h20kWzmcSTufRR3PW0LbvqdGP4eH0JteuCPsJnSafddEgC320U6IPgUJmeMmPDtEqAyTkarN6M2EGlGMcAB+xg1W3htSD6+LkX81ZZxX3hfjUcjenEbFGQT84FSUMKAOJP9HGamTBghDokJ0q+D4kW3LW7m/IxxMd0i6POWdxEBNw1VlKG9MKHtbcpf2I24Zjfiuh02AWgUJscYC796c4AJPJvdzpSC/tzsW/YFeEG4XEky8l2LBz9aBBg+wKhglaG142jOBzJJSxvGxsGTE3IE7yS7gAxdNAoWv2seHsaGqbpvPOUPSUVz1sKyi+orbS3lwoDiFtIK4Jq0o+Cp4bFTgx5Hn1Rqge3xzo8qRW/V7Df0vWmhc9NWuAMP83ALH7ob9+YTNZsxwlpNk9LpN6s2Gz6m7lsGB12GnZlyiwkquXR8nd+T4NE0Xk1brL8WvznqDibZcVzzyKHP+Jrv6PInrndnbsBmQVuYxgVHHAU+xbgGMzazNbA3Dnl6RI3zLMPKPyk77/H3anI9I+G0Hbklv8yMEwu3PLvpbwjWhDrFYcpnTaJAL2tCEz2PNvn3C5/P8tI= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Append rows to a dataset. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/attach-a-file-to-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/attach-a-file-to-a-vector-store.api.mdx new file mode 100644 index 0000000..4a0b251 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/attach-a-file-to-a-vector-store.api.mdx @@ -0,0 +1,68 @@ +--- +id: attach-a-file-to-a-vector-store +title: "Attach a file to a vector store." +description: "Attach a file to a vector store." +sidebar_label: "Attach a file to a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztWm1v1DgQ/itWPoHUbhcEB/Tb8iaqA1q15aQTRStv4t0YHDvYTtu9av/7zYyTbLLJtlvEh+spSMA268zrM2M/nt5EVrjcaCdcdHgTPR2P8b9EuNjK3Eujo8Nowv4SsTf2DP6K91KJ49l3eMCsyOFlob3UC+ZTwbj3PE5FwuawaBTtRbHRHr5HkTzPlYw5ijz47lDuTeRgccbxk1/mAjQZEgwv5tbkwnoZrJJJY43zFvTBmraRX7T8WQgmE7RnLoVlc2PJKjQmWu1VwnslzXmh4KvokhydOvR0RC9u6il9RxENZXuMqyu+dOyiK+IiQu0QGytnhQ8ebfrLk0SiAq5OWp4bLY7n0eHX+g1dKAXi6p9nxijBdfORLrKZsM0npaONJ9xavmw+KC1ZfVttOvynWO5fclVQeksXGHfOxJJ7yPWV9GkrznFa6B+gD0JgYcVi2UXUWfkNKxyiBRLlAB2+BhKKYlJ7w0iYgwh1IrEFK+Hbbo4Big4zzAtvWimvH/QaiGIaycXFFxFZjB8zwHPMKoej1WpbJudcObEHtfazkFYAnL8GA7/tRV56hZY2auxNKbCyYtJn4qSjn1UBJ/sCDhnhkALqRj0Jv2/4nEeVrQA2Hu0SwrC8DGL4oRHBvUpcBzNvjJ7LRWGphbCcW54JL6yr63xDVh0NsOx2n+mNqbkUVvF86s0PoZtFCjgUCyiohs/PxuNNdz9T1TEzZ0EA/MdKkWwm/JUQmvHkO4+hZVSoBm8zfj0N+p38R+yk/GVX+Sd+LbMiY3rTCPAyKNtjWeE8WFIb82Q8Zlwn7Nn41R/3Q25vvLa5shvCzyh3IcW/UEY1au6jrQexW8qJWlxc4W8GvWkNPygqUJpIFJRJzUEfpq9E2PIzLAPRpZUZ7IIIc9wPsabv3PmGbnZrN2s0i6GtDW1taGu/s62hvTtopSOfFXgWnHK/JcdNe85lJsCvLGdXKSSsPu9dcThXJgkcCAFj+LRZ86hFceenwtrQYdtCHx3nIa6P2TtcAadHKKAsVJWcBw1QH7FwDoMw5/AgubuGTCJ6z+HdDiIsIKc0r3v2Xi/EqE0V7BXgy3UsBDgctpCWP8EJVA+eJEScyrMxMQ8oBfSgCIGBeDq+6G1wbakfiozrfUhWQukmY1n5NgtLZ/UZvJJ/rxrCeK0N6sctUsiPkMt3ZawI5YXr6Y2FtdhXGnkLS6kVVJSjhxt0gh6bLFcA7OTW1Eg9BVULILXu1nUx17FQ6g5pJcRW24NwFvyGCBQYruls2aaHPV2y0yNRGGbP5dCDA5+aweaUSlexKBbEgpYmNZ3uQqqP3laRbu2/SOu51AEppaL74URi6a3Z75ob99HHVn+p0dIOWte57WE/rvRuMPtc6MlRedfBaDXD5SzYSR0RfHzWd0VyDiFC/6CxUR+TGiizTBiWF1fYiajZ/Kb7kHW93NFqP5yfn1QlQ6W5bul3pT40IFoMpxzmUmM9c0WWcbusQBH6x1Uq45TJ4LSVHAqWTpS6/J700LUCgEbtqjmsRtXKaPCNpe3e1XinZQ5qklCAWKJ362psGxP25fRj6QwUOJ4pqJpgM7IC3haXgmWIiebOwmem8OGYlotYzmGfNXFMbSsWG3bdpz5qjIds1cFrgDp0z/7gVXd6bG5NRjZMTo5G7NQUi1ThgVopc+XY6fs37MXL8Qs6Sotrjj2yiS86D1b6XvOEnQaIr+3ZEfll5Tx91V85sZLY5lN4Ha8TIeQGXtbLSjBKBSAs5CWeNjNTwBqMLRwkhqIaiuqBFdXTV2t954D0T4j0srJcs7T+NgXUBESoOieSUtwUGZ0fR+xECQ4GeUAPX8COzBR8aaud6vm2nSqcVRmEE0tJQHgQWIUW1xBvvNMVZRiG0hpK6wGV1vPmfnWEyAbL2VlA+1p/WV6TLuSrYCQjdlwAEgTPaFea4c2CNjTrSaryqs/knXnZLYKHohqK6mEV1bhP2w4lRFWCUOMLh94EXnV0HHXuOiY0NoYkE2GFUPMW3cRB8vpiimi+Drfqm5QPkw6Pc+7T7q0T+L+FzqLGYEJ9F+UNal0nw9sCktMpzWqwihy/PKu+Nsky3Bz9lipHa3ai6m3/6kiWfnUv1EYPeiiN3v7oG0yTbxtD6b2egsbmBN8BosCI0Y5ja+IqnWtUCG6BVdb4dYPRMLAeBtbDZGeY7Pw3JzvDwHoYWA9tbWhr/7O2tvPA+j5+VQfwhkc4KOIysBYcEJ2bhrbqipysaRMImtT61CTIUQxdoxNVOYwOLp8cGJKKn5q0xh3cbLCc1QF1E4w1XWwERlRY5GOp9/nhwQHXy/0UFIhkXyme8X3ISvxjFJuMer8TwNCkX9KLbys4f/2G3yE/OV0zmXdrPlgzkXUzatGHbUfo6nCPTXRFvHwe9pEymh/RQnaGFrKzkj9Tb6GZnnUBE5dPehgPXje4FvHmzauApmSfcpojX0oQcqHZxh/OnKCZgtBJbqDYHdUkSJE4hbeBSyeIRuwEdg79wwWh3IoLHIcqIh6wP3WEz3A2ogQ4g4PaYNUnkwgVtgSEQsaJFZa8dgdC3ApFg23u8m6ZES+u/UGuuNQ0gbZ0KxMA+TXEO0AS8/BkY8CK8Dvs0u+ATIARog/F3NzMuBNfrFqt8DHgyiLs4CPdEs0QAQBCOI3g56SuwK3uPTotS+ox+xVG3+t6hVCNWxQxSvgJPgLB7LlkIDaaCp7AboC2h1VvgoX756ENVlI61B8JTXhjEsci97eu/dZoGCfHZ+eweFZeL2T0iymR5Vd4TwH/ksWGIkb7Fj27iRTXi4J+NyQKMvHPv7nGA3c= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Attach a file to a vector store. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/batch-inference-coming-soon.tag.mdx b/versioned_docs/version-v0.2.23/api/batch-inference-coming-soon.tag.mdx new file mode 100644 index 0000000..e430bd3 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/batch-inference-coming-soon.tag.mdx @@ -0,0 +1,22 @@ +--- +id: batch-inference-coming-soon +title: "Batch inference API for generating completions and chat completions." +description: "Batch inference API for generating completions and chat completions." +custom_edit_url: null +--- + + + +This is an asynchronous API. If the request is successful, the response will be a job which can be polled for completion. + +NOTE: This API is not yet implemented and is subject to change in concert with other asynchronous APIs +including (post-training, evals, etc). + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/benchmarks.tag.mdx b/versioned_docs/version-v0.2.23/api/benchmarks.tag.mdx new file mode 100644 index 0000000..41402bc --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/benchmarks.tag.mdx @@ -0,0 +1,19 @@ +--- +id: benchmarks +title: "Benchmarks" +description: "Benchmarks" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/cancel-a-job.api.mdx b/versioned_docs/version-v0.2.23/api/cancel-a-job.api.mdx new file mode 100644 index 0000000..5ccc0fb --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/cancel-a-job.api.mdx @@ -0,0 +1,68 @@ +--- +id: cancel-a-job +title: "Cancel a job." +description: "Cancel a job." +sidebar_label: "Cancel a job." +hide_title: true +hide_table_of_contents: true +api: eJztV11PGzkU/SvWPLUSJBS16jZvtGS1aKlAIX1YUbRyPDcZg8eetT2BKMp/77meyReBilb7gpS8zJfv17nnXDvzzFOonA0Ust48Oz464ktOQXldRe1s1ssu/s4WB9n7pz4NCxKe/qspRHEvg9B2Ko3OhfOilGbsfEl5dpApZyPZyPayqoxWku27t4GdzLOgCiol38VZRXDrRrekIgwr7yryUTfZhShjHTbWabidkMfC7bT+Gg4vRbNaKJcT5x91NLRhG6LXdrJj2vceyafFB0KKUDgfRajLUvqZcGMRUTGlNfeFVoXQTdFeSxsFChbStt9TnAX7j1Kbl0ZuVnNo4yxqE0VdSnvoSeZyZEhs2Gylw5G0Rc1WvaDKNxfpTpq34kR8G5y3xSgkPyJRB8pFdOgsrGlKonSeUCb3M3VOyJGrYwoeKlJ6rJVwStXeE8I/yguJyTzXTbzLjYaOpQl0kDF/tAdPetfLDi+7tQLvZtW/BqZnwFtyWYy9K1MOJ5dnHTFw9aQwM7THGHcfxODPL+LjH0cfOwwaPciyapix5Beovo73WeZi0FB8nc8Lmb9YcPXvjz89rRxlNFQhCpgHvonOwdjOlo7ZK4gw0VPCtXQ11jC2uqS9qPaiemWiOv60jjcE078y01tlhU1p/eNqaAII0YMiyhk2FpuMJIwudeyIS0MSCUWwR04kRGLw0XdavX14bqcK5KdoPuBkKRHgYWLVlh6Ad8QTtTDspbWX1iuS1ofN/eqMmY3MxVXD9nX8Vl4nu5RfgpF3xEUNJpAs0640Imw81kUAhm+tvHIay9rEXYn9zPFeVHtRvS5RHT0V7QUSSiphqslJ4Gr6OBdy6tupfuEWGTT41o06zHHpZUlQLtvMM4sHrBoByQKku/tXs4I0W1YyFjuV8+52drrEfGWW+lfbphHIo24a5yyHXEMefY0W7AiwpQ2KWSWEbH8tFRhwEiqV+2tRgRkQKRyWIoQBOAknhOxl3ek7aapCdrmq7qre0J1vQrboIjzeNVkvYN6cABqMa8+tLGKset0uziKHhQto56ExspSHoIG66yhXZpxIIDRXx1kyPF0OwOsb/saSSPm3bDlne3HF9uKqFUbCHfE5eAPT9N0TyPEcCVuKkpsa3/QcCxkFJuNUw8l3Kx79MLQo/Vkgm1cOIzJgJqWDlPaYhr4RSQ6kBc9PP5aKQuNUevpumejOpwmw63zEf3oMZoKXk2VWXzFhTUhiqoAjxhVj0tLmMdu36p6vd4adhS0nIj3EbmVw0mP/qXHzlgrXWUsFrGYy4LKmAx56jzTEjODXLZPRP247+5nPRzhUfvNmseDXOJd67jdu02QfcXPR/VwHvs9Xo+vZYt4MWq6/Fb+jzydrb1+CrswmWPATbu9o9nhesGz/h2R3FPwbebVgL26wssAmht2aoWw+nihFVdww2zkdbI2C0/55f9jHfPgBFLd/Ww== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Cancel a job. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/cancel-a-training-job.api.mdx b/versioned_docs/version-v0.2.23/api/cancel-a-training-job.api.mdx new file mode 100644 index 0000000..cc32467 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/cancel-a-training-job.api.mdx @@ -0,0 +1,68 @@ +--- +id: cancel-a-training-job +title: "Cancel a training job." +description: "Cancel a training job." +sidebar_label: "Cancel a training job." +hide_title: true +hide_table_of_contents: true +api: eJztV01v2zgQ/SuETi3gjzRo0a1vadrFZrdFjMQ5LNKgGFFjiylFaknKiWH4v+8MKcVKnaRZYC8B7IMtiZzP995QXmcOfW2NR59N1tnhwQH/FOilU3VQ1mST7PSvbDPI3j60NCtROPynQR/EDXihzBK0KoR1ogI9t67CIhtk0pqAJrA91LVWEth+fO3ZyTrzssQK+CqsaiS3Nr9GGciwdrZGF1TKzgcIje/tU+R2gY423k/rj9lsKtJuIW2BnH9QQWPP1genzGLH9LNzlHzcPBAgfGldEL6pKnArYeciUMUY99yUSpZCpaKdAhMEFSzAtOsxzob9B1D6uZHTbg6traHaRNlUYIYOoYBco+jZ3EuHIylDNRv5jCpfncYr0K/Fkbg4+9IWIyn5HEXjsRDBErJkjUsUlXVIZTKeETkBuW1CDO5rlGqupLBSNs4hhf8pL0oMikKleNMeoHPQHgcZ80c54snkskO4Q+uueVd3+KU2PdK8jsti7mwVczianozEmW0WpV4RPFrbGy/Ofj8W7387eD/ipuEtVHViRscvovo23kcoxFmi+DafZzJ/s+Hq3x5+eFg5UitShSjJ3PNFsJaMzapzzF6JCAu1RPqtbEN7uLeqwr2o9qJ6YaI6/LCNNyOmf2Wmt8ryfWn9bRvSBHUIbyViwW1jsUFAoVWlwkhMNQIlFIg9sAASiaZFN2r19u6xk8qjWxL41E6WElJ7mFiNwVvqd6A7bNuwl9ZeWi9IWu/659UJM5syF+eJ7dv4rbyOdinfNaMYidOGmIBQxVMpRzp4jA3UMFpr5VXgHBoddiX2lOO9qPaielmiOngo2jMkFFXCVIOF52qm1oeZo0OKIBCvjm3Fv+fWmtdc0P0Cjhk4TbCHzuDa5iOWADiokITNLq9Su+jY/GiLFef8P0mLgn1vGlX8mjp8ml5cnHzqUCZL5oqM+Y/+G9x3UXv4pkZ0bfvT5t0beGzt1jq4BukBtaa0dJvVNr6l1xBKuhsv34CuSxjz42HX0zEFHKdMaWt6J+C2rrPGMbhlCPVkPKa3k2FJdlgMtYYKhkQM+WMkbZVtrtiO4FZhFQ0/dSPx8orXuO9nW4Q+b8m17XDX2ajYuY0tb4v/wuHEOYcT562yIqCULueaIFi+eQAVHkT+niShPyT6nkMJQRD+S0VOvhnx04emHsZ/G2iK2tKM9TTU4puYcjROXVJZwRzlAezmINEnp+Dwm2GlWBdHyK7znP81aRoqDhZdVl9pRGsf1chw0bzjnhii/VPCuNeAnhAet2iJHfA2jGtNaxwxIr9ueXOZtbzJUip3zKF7csJHWWIPQc0MYYv1Oqc30gunNxt+TNi7VZJqPBZyBpbIQQY0w+mwYtb8wBUnmlIezjgv3q4bzm9Hx5tBZ3EkJdbhyb1XPU1MT89ntDlvZ0XFR+Ekc3BDD/l7ktGFjS2MUyA+W2cazKIheGg9+eTPv8I/V7c= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Cancel a training job. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/classifies-if-text-and-or-image-inputs-are-potentially-harmful.api.mdx b/versioned_docs/version-v0.2.23/api/classifies-if-text-and-or-image-inputs-are-potentially-harmful.api.mdx new file mode 100644 index 0000000..cb13f68 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/classifies-if-text-and-or-image-inputs-are-potentially-harmful.api.mdx @@ -0,0 +1,68 @@ +--- +id: classifies-if-text-and-or-image-inputs-are-potentially-harmful +title: "Classifies if text and/or image inputs are potentially harmful." +description: "Classifies if text and/or image inputs are potentially harmful." +sidebar_label: "Classifies if text and/or image inputs are potentially harmful." +hide_title: true +hide_table_of_contents: true +api: eJztWN9PGzkQ/lesfSoSEFpd1StvtNfTVWoFClSnE0XI2Z1k3Xrtre1NiFD+9/vG3rCbH9BQ8VIJHkKya4+/GX/fjD23mSNfW+PJZ8e32aujI/5XkM+dqoOyJjvOTkRlC3KSfwo7+kZ5OMz2s9yaQCbweFnXWuVxwOCb50m3mc9LqiR/C/OaYCbNxMTa2ZpcUGlJVfTG+OCUmWDMKoKLkkRj1I+GhCqwphorcmJsnQh400PnCGM84C32M36sd7Mdh4rGUyGCFRMybI82jftGBx+Nt9975qVzcg7rKlDlf+72WMvJhPq+j6zVJM0GwH9LAhAnpJkLO46gRqTtTCDgNLEOBoV0JFqLEV73ahsSWRSKjUt9toJpDchiHcmJ0MqHJYhujX1gK8SsxYmPeR+QwDYZG/q45teRMVRcK1M34ZoX/hWk98W83erFzh4Iqa2ZiJkKZXwTYQm29sLv4YkM8bHPLfxK2D2osupTfPsrbpimGpHbPd5raJVLwPDci9pRofKAsI/midcRJLjtrivyXk5oS6CgFgqykGGrYO+Dbw2djrPjy54jWsPYFiZt+rqJYW1bew9aJIurBW/pdjhjqT2xMH80yrGwLu801vPuaj8LKmg2+vlO2afR/LCV9OYubEl/D+zVxmifPQq1ioBj8uryzAO4dwQcQ/fHtvzOGbBNnGIGCikzlVpF3VZSI8tWMYZPlO99kKHps1/B7ASUWHfjn4uLM5FGixwOcczbGPwsp39wjmsDD0ZuEr60LgjfVJV0dzmU4phZqfJSqOS0U9KEWFakad/HdeJmB6l2qCZp5TSal2ahIieWTSXNgSOQcKRJ9OaswOGVlIHPJt/ByxendSLUnjgRX4afWmdygB/RXTVzhNk05UrmOLHxfiZqyJFt2rxWU46amgub541zhOXXcD2GwO0OL3frLng9Dscw3RO85YFEjJ2tIoaTs4+HYmibSann2B6N4ufF8O/34s2fR2+iFulGVnVixpJfoHq33jtZiGGieIdnR+a3ynn1drtyctQCsKbEdM9fgrWYjFrdGmarIMJETQn/K9uYlNFVRc+iehbVbyaqV2+79S7A9M/M9FZZvi+t/2wDTSBCdJMTFRw2Fhsfq7WqVDgUZzgaAFAAe+REQiQaL92yUr2+r1LhIDPF5iOcLCVCeJhYjaEbxJvPPdSG4Vlaz9L6jaT1ul+vPjKzgVycJ7Z367fyOtmk/DIYxaE4bcAEklWsSiNC4cEFjC/NxVJeBY0lDpZbLvsPGH4W1bOofi9RHW1bbQcJRZUw1eTEszfnckxhzuBXwb7X0nsWFriDUNFN4FbIAJZUhZt26iOk9kxtWThKasAvpavGjeYuWi2dxO2UHK9zlWKIWvrOFnN25KmabIxj65X9vtv3A02VjSh8jN2SF+x19HePKZen0MwPxfvERUgQ80H8ZIibRiKuxSxLz/y+SLq7e14hSakD3GeRC1NTpr1Tw1iltHS8lI19p3hl9o9t/bUB7l+ZUzdwjgPMzDa6wHnlO/EyENPhI6/yMexLPD3qDxvTXeOXF5PIuG56cA2lxkxp8TOrbby81DKU+DWYvhxgbSMVf+vQs+TSEcnHbW4cc70MoT4eDHBYOyhhhooDrWUlD6CT/PthbquMd9UT2K/Ac57417JCXF7xO2bcsOPmh05rLbe6ILfR7zWXOOHE7Wjd/8SLi3NeXJy3aSeix2xGnrZn+nLLjnGW9iv5SvYzaN9ybNtBB1MFI1+NWPsDHylexcgUtUUB8rGNmRpquXUpBRVwQXB1cmOZc9OPjULPXw2nEetift00PuIrpUbGdZwGEqrPHUF5L1EMOCYG8n+STLISqV7meALTrYx46qDWOKvHlqKLEkt8vEy7lRjJu/iyJcKSlaAQM49H3t6OcPD/4vRiwY/BKTdPyS9W3xFTBKTDBJRKnAmYjd9pzp4knw4uGA8P103MWOuZkbNZmnGS51SHB8de9TR2dnp+gcGjNvuyA3jq5Ixbcfg8zqKTySUMiM9uMy3NpImd1SzZ5L//ARJJ5iE= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Classifies if text and/or image inputs are potentially harmful. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/create-a-new-open-ai-response.api.mdx b/versioned_docs/version-v0.2.23/api/create-a-new-open-ai-response.api.mdx new file mode 100644 index 0000000..b3073b2 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/create-a-new-open-ai-response.api.mdx @@ -0,0 +1,68 @@ +--- +id: create-a-new-open-ai-response +title: "Create a new OpenAI response." +description: "Create a new OpenAI response." +sidebar_label: "Create a new OpenAI response." +hide_title: true +hide_table_of_contents: true +api: eJztXVlz3EaS/isIvowmgqJl707Mrt840iiGO5alEOXY2LAUbbC7SGLcje7FQYrr0H/fzDqAKgCFq5skuvubhxFNFrKuvCorv8o/ThKRbtZxKtKTH/84+eHVK/5nIdJ5Em2yaB2f/HhyHgfvNyI+v/iom76/+peYZ2cnpyfzdZyJOONvws1mGc1D/ua7f6X84R8n6fxWrEL+KXvYCCK1ll/Sh5tkvRFJFqlu54kIM7GYhZnVNiLSNyKhxu5wfomjr0EWrUSahatNcH8r4iC7FYGZSHAfpoGmePLt9EQkyTqpz+rFe/lTuPxz8HduESxEFkbLNIiuXXI3IhaJnFdwTQ2I6GnXdNYLYU0kzZIovqnNQ/XKbYNoQYsYXT9QM9k3fxisr2V/eSJ4FjTdNLzpQfYf+SqMX9L0F+HVUgRy9oH+OlBNr0w/BX3qIFwsIrUgH6zJXIfLVJwSk/xvHiU09R9/VbMrB/SFViPKljwil0vk/Hjo0aJ71LSp/5sXCxGJJLimYWe3UVpshFwF6nvZTe0dN7OJ5alYSIrlZjI5vX8N9IixU/pDIR2yi+swX9Z+aferBENtX9n7aRAu78OHNPhcfPn5RHafZ5vc7j5MkvCBqEaZWElOWsfi/TUteqcAlXJY+0TP6dvp9t1k4mvjarmr8Ik5mFoGelTMycxtUUyzNazI81dk/IsvP5jJTu3lr/za7vq17tG/A+XHvAdD+F53KAfrZfoLJv9OTVEP5lPTQD/Z68OM6SwOKSGjc0s9RK1WIWldayM926Q0WRszlGu8XN/bFGt/v41ublsbhHm2Pvn2xd4g+avqjH8SdySTxApqdGrSK1ZKNPg5zZvIngbzMA6uBO0TjevzySn9wAPgn6j5Z0lZyU4/3pEdNDBP+fsx3CO/VsOQP87ypIdSsgzOLx9/KoRCLoGR30EMqbd5BEteNE7/wh7KUKZkBoiY2iqKw0wZXM2VDz+HKx6WHCbpcPIUeH2ogSXJnT4CdM8kdI8jQJ2bBk0ETfTUmohH3rt/nqffLxqjkXpwhPL76srF/j0vfxyvM+mrprvwEbtGdU1nm9k8Uj0646r/xT0emmG2sKtDQjGs/NW4s8HaHNGuRSLiOfv2RMwQjaW56aLKRqmFUBQvxNcep9EP61TKSCA/MATNTIP7iM4w6nw6SrK0zTRrZU3QDNErbeW+vKVvXvu2j/9YjrfkOSl2ZmH4wJhuxJyWfy6XSIpgIXt6an180y4uJAXWzIS1PwzmQZuCYkERL2Z9N/rv8SLYmM2ubnO6CZkD3H0+pdmFSda7h0tuPbgPvfmdjhE3a+D3e3HF+7jOk7nk+14GxLIaXlIjuLzcDnfpzCTV6HowvJfZeeB9eJ00sEho3M6c0l3wN+9cGMUimfn1ra/Nt1PrT02as4Ol25Ruq+7sZOUxu+1Mxd38RnVnj6APDxjyjvbbfgPl2DZhdls3kfK3HYu8yxUsl2mQKfgghznqwObyY7eHBJcDLsejuxyua7A1U8IDgQcyZQ/E6x5szfnwTZ7ZN7HciN3Y1gPwU/zhnHMrPvJtdDjVjbN4h/Zexmbc4JH6nYyxfvv2hQaQrJUi7A52pg9pJlat4cwFhyl5Gq2t8rSjQZimEbEnq+ovPWJTJkRuM40VNvdwC3WQ5WnDn4ZdLCuTotexK5r4zhqpGztNlJewSINsLU3EXZhE6zwN9CfSQqfGkhl6aXD+4eIs+HQrHoIwEWS1l0FO/Eq+X6yv5K/EPKQFr38W3ER39BP9fiW/4wYpe3uf5SQ+n9AQlrkgV4BsOP2RyEdpEK8DcgzkIIX6iL7esAmmoc1ppUhHBOmclAwNv9fpc8ur9my9XgZzmkDbntbWO6f1jrNAtTdWmc1WKsJkfhvwAAsN18V/9N1MfTeTA7H5sOlvjodhht/iglVoDL2IkVpMr0wXgzoq47/F1aXslgf5umn0/12uWLEPgQoJF/kbvFOVwPeT8wU1Shwb5I2bl/qhcgFDCon5RM9WEySXR8xznbSzHfOxwRnFfdJS+div8Y9D+a9KRB0DaDOpjz5r6tngMKOpXNHqNbZtZu8GOxXnco+L/76iGYnQCZvE+eqq0eTULlLqbCktkPcu6p/i4aVUk0E5m4DM13oecS6XPNgW27ub47xkFO3smaQotRfjT/btNNP5OrEJ6uWsEvwoluIupENNID8oRVFztSIYvLgS2b0QcfBKWpbv/yyZvN9NecMtuUN8mGK0GNDnI8uZ6+H1U5rsBbpa86MWlNrJ1R4554xleRITz1w9+BXCWV0zWdw4huIIW2K06TirUl+g5mDLY9kVpjVSAhVHU9d5PJcnYGNfRkicIXElWOCYkLIiYXKTr6jHHobkvy7f/xyov1WFtyBfkutjSvRXDXak9pfBRsSmoK//e2yCxd3nBZM27ckoP9Ci7rHK9k5raz9YaAzHaT6x93iY5OjB+OXGjHZqzti71x+Kvek8zc03df5zfjmU9czHiut2I2BmRq6ADdcDTGclstv1oqYJ6JR8J5LZMrzqkz/8EzerZWUzeUUnuCVruzS/N5tRJJr3lpO/O8nZ0bW7GjrbvClRuJu0YnbjJ1wn65Wy8PmcM5yu86XFR4ONlgmcWKJnjLy9zv0kkcbRKIAqiVuGWshToWFl6znx5wtq/+epyeOSTzSsKNJh5w2WJv52Jr+tSWnlT2NktSShJHbHgkAreRctCtyCHCrPW/4w+jSjktv86JEJn2eUmpMjD2gzSyXHS/KnVCeybcKE5CUTyUhNx8Rq3usA7fCm/AOTvL8Ns9LwL9YiHagU7P3S87GEn/iFj/zMsin/XzMzq9WSQ7KdbZ94NXjvJq4Q3pHmlOgX9Y1i0tJ3L/l3vO5z5MgwfG+FVyxHXetVpuuqOMtwVicpVXwYW5Mbm5VdhxoBdYLMb6BOkOs9uVxvoE6OQ/cAdQJNNG1NBNQJUkAHHuKQAgrUSddGI+cTOZ9Anex/ZieyOYE6gctxSC4HUCfwQI7IAwHq5GB9E6BOgDoB6gSoEy/qpAa86E7gAA7lKHEodYzEk7IKoCmApgCaAmgKoCmHD02pwCi6k9UAVfGaFkBVjhSqUmJBHttPA2oFqBWgVoaKpgUAeQoBBYwFMBZb7wHGAhjLFGAs/mi8GpF/DXW9F1o0PQQpusELk9N5WvpM9LPI5mfynMqMTUZ4KVWO9BxscTfiZ+k5vcyVSNqt4Kir1YXJ7zXxJA65mr5kv4m445jxzJiCXicWSyIu3hjZMqTKy2vqK+SlpzVNC9W+XTSroUiSOuavJH/nfeIJ9sE2XG2Ug1IS8NXuMaGEhtCBSsrNtM92Hd3kSXndaQ/byIFXgytK/a/Fuq9gdJqtvwHXzjI6qLNdocBr8pcT/8tEcpX4z5nkFuUiodyi8vmE7wSMkVTHIDl9aZr7K3gGQ8ROzombKx28j5cP5a5aozpTUag9NJc8Z9tkOvNOb9f5ciFZkVaAFvksuGBR/PBATn8cXL7556lywSKWq0zw337bPCxC2oj5b4GsedW1amMt6HmwcG3omB2jXpwqWrZ+9PRbqsZAfU4dCF6fubrnshaTVouGlfA9ljO4FbM4DW+uIq965cXXkIm1j3nwbXJNvqypWCqnom5UIpH0N7X10c00fX1o7BpNs+X7pPP1s/VmthmkZX/O50tBOjw12rZw47y6NsnjeTXdo5u5PhWf8SaTFb55CGSZQOpE35Haxd3kpfIgGAl9YJ9lmgLyvNQiHRgxtuoRnip3SRWeOy2VkT7cNrsKdsS0PaRsyFUvk2lnaENqcBPVve0/WdUJTTyYw+20neEizDSz8yC+E3e0UC9pVUW44mUuFW1fLEexWd3BVhR0REFHFHREQUfAWwGtB6B1HwGto5AagNbvne4BtB6aaNqaCNB64NwGXiwB5wZofddGA9gGYBug9fsPX9uJmTwAyNqoAxug9XA5JudyAFoPD+SIPBBA6w/WNwG0HtB6QOsBrUdBRwDpp1SzBKh5oObLJkDNt57sgZoHar6/LZk2ar4z7wwoea8pAUr+SFHyj+yMARoPaDyg8dOSRyDhgYS31RyQ8EDCTwEJP+qSHwUdkfkN1Alyvfcg1xuok+PQPUCdQBNNWxMBdYIU0IGHOKSAAnXStdHI+UTOJ1An+5/ZiWxOoE7gchySywHUCTyQI/JAgDo5WN8EqBOgToA6AeoEBR0tjgQOBQUdAU0BNAXQFEBTAE2ZIDQFBR0BVQFUZUuoCgo6jpc1oFaAWkFBR8BYWnYEMBbAWABjQUFHFHRsiSegoKPTAAUd99Bc8pxR0LFJP3r6RUFHFHR0NPE3FHS0L5N3WNCx+/hmyJ+Zqon2Aa7xj06C2F07ZqhKYOhVZ9VYdyzmpaxYaX732jNo1UqmPMnhR/EiYlZljzSIxX255LdhSoaUlLIefp907EHeW0vWp5V1WASeIuNNy1Aqp3MMQOMAsW0E7BhQk13bBJwkcJJAbEP3ALENTXR8mgiIbcCnBt5XAD4FxHbXRgMvBbwUENv7j4raiZk8ACTUqAMbENtwOSbncgCxDQ/kiDwQILYP1jcBYhuIbSC2gdhGnUDgs6dUCgNgbICxyyYAY7ee7AHGBhi7vy2ZNhi7M+8M4GuvKQH4+kjB14/sjAFxDcQ1ENfTkkcArAGwttUcANYAWE8BYD3qkh91ApH5DdQJcr33INcbqJPj0D1AnUATTVsTAXWCFNCBhzikgAJ10rXRyPlEzidQJ/uf2YlsTqBO4HIckssB1Ak8kCPyQIA6OVjfBKgToE6AOgHqBHUCLY4EDgV1AgFNATQF0BRAUwBNmSA0BXUCAVUBVGVLqArqBI6XNaBWgFpBnUDAWFp2BDAWwFgAY3mCOoFNCdlWYUBaLuKM+5DO7AteQFMm0KoSaBUJ1AmRfa8vL+R9uHt9yUW/uF8d7NVD4QVRyouUMB2gZ1poe1yRyi+yiHxg9Y3U2+tkIaRnkrpVb3o5/UX5HjNdGu+ZXB5Hf3c0c26K+xYMqpEaWzpIX+/wnUZl2+qLPKbKkOKuC6J/3jjjy0q1Id6U+1tZ6I1rDdkcGBneqxTIQtkhAMD2BITRtU2AXQB2AQAYdA8AYNBEx6eJAABDNvbA8AeysQEA69popF8j/RoAsP1PskZiNQBgcDkOyeUAAAweyBF5IACAHaxvAgAYAGAAgAEAhrJDgHtN6WVtYLuA7SqbANvVerIHtgvYrv62ZNrYrs68M2C5vKYEWK4jxXI9sjMGABcAXABwTUsegdcCXstWc8BrAa81BbzWqEt+lB1C5jdQJ8j13oNcb6BOjkP3AHUCTTRtTQTUCVJABx7ikAIK1EnXRiPnEzmfQJ3sf2YnsjmBOoHLcUguB1An8ECOyAMB6uRgfROgToA6AeoEqBOUHbI4EjgUlB0CNAXQFEBTAE0BNGWC0BSUHQJUBVCVLaEqKDs0XtaAWgFqBWWHAGNp2RHAWABjAYzlmcoOzderzVLw6dwu/4KKQ55SQgvSOo7mbm/lXBGPqTfElKZfbuhN03wvvdWG4mqpoYIJe72moEIeY9nPSsGwwQhKwSyzsFu1XsTzRLBvRMzm4BmUY6qqTX1TK79ddMteJkU83yzCzHYct5JCsXdCyOt9pvapTQorzbYRw5LUUDl0GdWwV8kXjyiUHBt80zj9JqmMfAytuK3fGyePJpX9oqFvIw59GEXSiDOypGk34tloO49cNLvso9tqe8EcYR+rcqkzDp5KLAfZSsnHhrsGGsoR5syN6BUxgN1bNrejI7dtTvR3Vqx6m53r/mSUaLWSHSpnT2LvTFSXw0znZrzjTF8z8w8xggNCphV7VYqaHUv1Xqns0nbViEP4JMt7LVnnF7sVvREWzg7ePoP0DbJwPpPDaV6DzN32MmFlQR2nLFTSrc6ieEZrfEN/T5tFoeuDUZLQQnRwAthjMn+RJMbcfxF/8E28kfVdVpMJj2aSWzF7N8/24rptuEb9bCSvk2fK5hPd2stygE+ghErDDHVU4ZRiafoxltt8J6qoIDldRfTaO+smPeTltl5aaLd6xL3B7DY+Le0H7c0utuDdfFNc81im4JmXUGcZ9Fk9k5DwnAv31iRFPOuidWgZb+tnXbpS6kfHhdqOlc/gVZhkpLNe8RBv6+lFLGjbGoIVWwUX9mDrvKdpX+NpHnhrmyeH+gRuoZ0ZeFxeYMEgvRyC3R1Dm6hNyu3TvDjw5GkYafCRc6Jc0eXjVBttxwuK0FA22OFuv22eSdNOq7GWG77Hm9zDJ9vVka9O6xl3e+xhbtCe2/lC4w2U8/pHkQAu7yyLq/zdZsR4OuJ8S9mb/Ff6u/Wkt7IlfRtmwX2Y6uvN05092tfzKUHPU4VDOE6nKbYDszQw/gPN2QbHb/3oUiKu8zRcViRT/U4iI9WPO5mhIdZnkh/NGKjx8NeU7E3qTGE+EE5o2KtD5Ql/rqzVenq+sEldYcV1ZvRVg0H0tBtlEOu0tk9IVT9JBb0bX9jatfPG6fqzUYNY3LsWIdK2gF9/CEvDJhNyYU9bk8idb2BKD01twpTClB6iKfWHJ5ubbW9IdwPseEw7OhDUUTOhg5J3zNS65WVOAyCiszDrwUBk974GWbQSaRauNjql1jaufO7TFB2YbwesVz2qnxpYb0HuRsQasRYUIaeO6awXjfxa4THZK7et4TMl45E15/7yRD6rUi9N5CH7j3wVxi9p+guJLRMOZFk1vTL9FPSHpVovpDHQA/Kyopxf31cPWuC5BRvxKjDGuJuagiJbxPKUkyKJYrmZMuiu9q+HamlUJLV+lfT10BsKyluDiY+uNWD8x6ZPfGhY1KPa15ow3ZfBqAKDKjAtLIl6VNA9qEcFTXSQmgj1qFAcosc5CcUhDGgN9ahQDQLVIFCP6jhqPuzETB5AnYdRBzbUo4LLMTmXA/Wo4IEckQeCelQH65ugHhXqUaEeFepReetRdSkFVJ86zupTT8kXKDWFUlMoNeU/2aPUFEpN9bcl0y411Zl3htJSXlOC0lJHWlrqkZ0x1JPy6QHUk0I9qeeRR5SPQvkoW82hfBTKR02hfNSoS/461AioE2R+A3WCXO/J5XoDdXIcugeoE2iiaWsioE6QAjrwEIcUUKBOujYaOZ/I+QTqZP8zO5HNCdQJXI5DcjmAOoEHckQeCFAnB+ubAHUC1AlQJ0CdeFEnNeBFdwIHcChHiUOpYySelFUATQE0BdAUQFMATTl8aEoFRtGdrAaoite0AKpypFCVEgvy2H4aUCtArQC1MlQ0LQDIUwgoYCyAsdh6DzAWwFimAGPxR+PViPxrqOu90KJZVfHS4IXJ6TwtfSb6WWTzM3lOZcYmI7yUKkd6Dra4G/Gz9Jxe5kok7VZw1NXqwuT3mniSrNet+5L9JuKOY8azQfUGLYm4eGNky5Cyah3qik60pmmh2reLZjUUSVLH/JXk77xPPME+2IarjXJQSgK+2j0mlNAQOlBJuZn22a6jmzwprzvtYRs58GpwRan/tVj3FUxWqSpYa/CvlA5LWgd1tisUeE3+cuJ/mUiuEv85k9yiXCSUW1Q+n/CdgDGS6hgkpy9Nc38Fz2CI2Mk5cXOlg/fx8qHcVWtUZyoKtYfmkudsm0xn3untOl/KMpq8ArTIZ8EFi+KHB3L64+DyzT9PlQsWsVxlsvDab5uHRUgbMf8tkDWvulZtrAU9DxauDR2zY9SLU0XL1o+efkvVGKjPqQPB6zNX91zWYtJq0bASvsdyBrdiFqfhzVXkVa+8+BoysfYxD75NrsmXNRVL5VTUjUokkv6mtj66maavD41do2m2fKYoZbbezDaDtOzP+XwpSIenRtsWbpxX1yZ5PK+me3Qz16fiM95kssI3D0G4oT5VGV5HE3MKTiqGHUR/oQ/ss0xTQJ6XWqQDI8ZWPcJT5S6pwnOnpTLSh9tmV8GOmLaHlA256mWyKrNYg5uo7m3/yapOaOLBHG6n7QwXYaYEtOv4ZlWz1OUdnSOc589Oklj/epmaxNg6md2hs8YKmL6BX1ZqX0bxImKWZc+0XPbbMCVjSorZrn85KpW1XAhdoLLTtSgm3h3hRhVNVNFEFU1U0QSmGO8ZAEW8jyjiUT4F3jPYO92D9wygiaatifCeAcCFA2/zAC7EewZdGw00IdCEeM9g/zGDOzGTB4ATHHVgw3sGcDkm53LgPQN4IEfkgeA9g4P1TfCeAd4zwHsGeM8AVTTxesGUCsXgqQI8VVA2wVMFrSd7PFWApwr625JpP1XQmXeGpwm8pgRPExzp0wSP7IzhPQK8R4D3CKYlj3h+AM8P2GoOzw/g+YEpPD8w6pIfVTSR+Q3UCXK99yDXG6iT49A9QJ1AE01bEwF1ghTQgYc4pIACddK10cj5RM4nUCf7n9mJbE6gTuByHJLLAdQJPJAj8kCAOjlY3wSoE6BOgDoB6gRVNC2OBA4FVTQBTQE0BdAUQFMATZkgNAVVNAFVAVRlS6gKqmiOlzWgVoBaQRVNwFhadgQwFsBYAGNBFU1U0WyJJ6CKptMAVTT30FzynFFFs0k/evpFFU1U0XQ08TdU0bQvk5+piqaummgf4Br/6CSI9a6gqQg8bf1Mz6Av26pnxuK+qYKmInWmL8XUnDR4hr3KM5rRkLKY26aEWimJRVQqMq62jLNyrscAqA7g3Eb6jgFS2bVNAFECRAk4N3QP4NzQRMeniQDnBrZq4GUGsFWAc3dtNMBUAFMBzr3/kKmdmMkDgEmNOrABzg2XY3IuB+Dc8ECOyAMBnPtgfRPAuQHnBpwbcG4UEQR4e0p1MoDUBlK7bAKkduvJHkhtILX725JpI7U7886AzPaaEiCzjxSZ/cjOGODYgGMDjj0teQT6GuhrW80BfQ309RTQ16Mu+VFEEJnfQJ0g13sPcr2BOjkO3QPUCTTRtDURUCdIAR14iEMKKFAnXRuNnE/kfAJ1sv+ZncjmBOoELschuRxAncADOSIPBKiTg/VNgDoB6gSoE6BOUETQ4kjgUFBEENAUQFMATQE0BdCUCUJTUEQQUBVAVbaEqqCI4HhZA2oFqBUUEQSMpWVHAGMBjAUwlicoItiUkG1VDaTlIs64D+nMzoVeihqCVglBq4KgKQzT8/ryQt6Hu9eXXBGM+9XBXj0UXhClvEgJ0wF6poW2xxWp/CKLyAdW30i9vU4WQnomqVsSp5fT31IHx9bfHc2cm+K+1YRqpMbWFdLXO3ynUdm2+iKPKUGkuOuC6J83zviyUoqIN+X+VlaB40JENgdGhvcq1bP8NYkWZAa6bQxKEgGg0c7YAIedAJJxjJCMUbmGAIftne4BOAyaaNqaCOAwZGoPDI0gUxvgsK6NRmo2UrMBDtv/BGwkXQMcBpfjkFwOgMPggRyRBwJw2MH6JgCHARwGcBjAYShJBCjYlF7dBu4LuK+yCXBfrSd74L6A++pvS6aN++rMOwPOy2tKgPM6UpzXIztjAHcB3AVw17TkEVguYLlsNQcsF7BcU8ByjbrkR0kiZH4DdYJc7z3I9Qbq5Dh0D1An0ETT1kRAnSAFdOAhDimgQJ10bTRyPpHzCdTJ/md2IpsTqBO4HIfkcgB1Ag/kiDwQoE4O1jcB6gSoE6BOgDpBSSKLI4FDQUkiQFMATQE0BdAUQFMmCE1BSSJAVQBV2RKqgpJE42UNqBWgVlCSCDCWlh0BjAUwFsBYnqkk0Xy92iwFn87t0jCoRuQpMyQr4Niau72Vc0U8phYRU5p+KaI3TfO99FYiiqtliAombCw9xD7CGfkLWaN5aARbjeVNKz/DRioo7eMOwKN3L+J5IthxIk50wA7Ka1Vlqr6pbdku9GWvoSKebxZhZnuVW4mo2DsJtRilTUQrzbaR0ZLUUCF1GdWwV8kXjyixHDh80zj9JpGNfAytuC1tEdle1cIeT2L7hVHfRhwzMRqoEaBkSdpuRLfR6B652HYZVrfV9kI7wrBWZVanKjyVyA4yspKPDXd5LawTd5wVZ/6+1naETXRjhkWPuzePbkdHbiDb97lR7Lo/GSWDrWSHCuSTGE0TN+ZA1rkZ7zj72cz8zZbUu1C9rOqA4G3FAJYiaUd1vZc7uzSGNeIQ0mLHh8noVmayjepQCbXDyM8gpYNMps80ccKZx35W8pPOonhGcnhDf+8T+NxadKy0reMUmbb1b5SYrg9GCUwL0cEZa48pI0VWGwvJRfzBN/FGCXFZTWZomkm2y4T6WYOURkpEN2P3Ys1tWKucRi/GKptPdP8vywG2bF6h955CnZWeABSbdx96cZ/bfCdKrSA5XZX22jvrJo3m5baKPnMvVoeZ+N2qpJaRNHJFS/tBO7iLjXo33xT3YJbpaVlonSLxvGusB9FneU1Kx3Ou7NsircQ3ygEK/VEXtkOheVs/6/KWCqa6wlJF7ip+1nasfgZPyDs/7841tp5eZIe2tiGo07W128Zf9mB3vQEHX+NpxgRq+yuH2rS9T3xut9M8j8uvbVzxdlbbyRG9idqkHFnNqwNP5YaRWo7jxcx35089C8N0eWLVRtuxiSI0lEN2yAhvm2fSxARqrCUvePZ/p57fs7BAD59xV6ffOq1n5IWx51oPR5hrbEYzn6kb106OsHPKxts9Q8UFCcjr6SJrY7eJUZ6OOCdX9ib/lZ54PTGybEnfhllwH6b6gvp0Zw879nxu0vOc5RB+1Kms7eA9/XjCB5qz/YDC1g9zJeI6T8NlRW7V7yRrqh93MkNDrM8kP5oxUOPhL27Zm9QpQQfCCQ17dag84c+ntlpPz8VuUO+N5tLTbpS5rNPaPmlZ/SQV9G5cbGvXzhun689YDmJx71qESNsCfiEkLA2bTNr2W9t+wYNjNLbGZ3G+gZ09NJ0KOws7e4h21h8ybW62vZXdDTLoMY3sQFRQzb56cq4GxDHMNz1QCDQqIjkLsx5cRcbwa5BFK5Fm4Wqj861ti8snRU3RAY93gMVVqYbUgMULcjci1jjIoAhwdUxnvWhk4grjyV65bQ31K7mRTDz3lycyal4veOUh+498FcYvafoLiVgUDhBeNb0y/RT0h+XhL6SF0APy8qecX9+3NFpA3wUb8Sowcr2bmgK4W8TylBNciWK5mTL6r/avh75p1C61fpVI9lAmCiBee3xgdAUL41Q2feLDWKPK2b5WGuq+2EZtIdQWamFJVDmD7kGVM2iig9REqHKGkiM9zkkoOWIAiKhyhhojqDGCKmfHUUlkJ2byAKqHjDqwocoZXI7JuRyocgYP5Ig8EFQ5O1jfBFXOUOUMVc5Q5cxb5axLKaCm2XHWNHtKvkABMxQwQwEz/8keBcxQwKy/LZl2AbPOvDMULPOaEhQsO9KCZY/sjKFKmU8PoEoZqpQ9jzyiKBmKktlqDkXJUJRsCkXJRl3y16FGQJ0g8xuoE+R6Ty7XG6iT49A9QJ1AE01bEwF1ghTQgYc4pIACddK10cj5RM4nUCf7n9mJbE6gTuByHJLLAdQJPJAj8kCAOjlY3wSoE6BOgDoB6sSLOqkBL7oTOIBDOUocSh0j8aSsAmgKoCmApgCaAmjK4UNTKjCK7mQ1QFW8pgVQlSOFqpRYkMf204BaAWoFqJWhomkBQJ5CQAFjAYzF1nuAsQDGMgUYiz8ar0bkX0Nd74UWzSqVlwYvTE7naekz0c8im5/JcyozNhnhpVQ50nOwxd2In6Xn9DJXImm3gqOuVhcmv9fEk2ThcN2X7DcRdxwzng0qQmhJxMUbI1uGlFUdUZd5ojVNC9W+XTSroUiSOuavJH/nfeIJ9sE2XG2Ug1IS8NXuMaGEhtCBSsrNtM92Hd3kSXndaQ/byIFXgytK/a/Fuq9gskqpwVqDf6V0WNI6qLNdocBr8pcT/8tEcpX4z5nkFuUiodyi8vmE7wSMkVTHIDl9aZr7K3gGQ8ROzombKx28j5cP5a5aozpTUag9NJc8Z9tkOvNOb9f5UtbW5BWgRT4LLlgUPzyQ0x8Hl2/+eapcsIjlKpPV2H7bPCxC2oj5b4GsedW1amMt6HmwcG3omB2jXpwqWrZ+9PRbqsZAfU4dCF6fubrnshaTVouGlfA9ljO4FbM4DW+uIq965cXXkIm1j3nwbXJNvqypWCqnom5UIpH0N7X10c00fX1o7BpNs+UzlSqz9Wa2GaRlf87nS0E6PDXatnDjvLo2yeN5Nd2jm7k+FZ/xJpMVvnkIwg31qQr3OpqYU3BSMewg+gt9YJ9lmgLyvNQiHRgxtuoRnip3SRWeOy2VkT7cNrsKdsS0PaRsyFUvk1WRxRrcRHVv+09WdUITD+ZwO21nuAgzJaBdx7eG4o72Ec7zZydJrH8RTU1ibPHM7tBZY1lM38AvKwUxo3gRMcuyZ1ou+22YkjElxWwXxWxxSq0hcDtq+O+vXjU4KyV/yqKVUXwXLqMFW+ZVuGRdIUdswayl9CiR+o5VGv/ObzBdH6Xm6fny8f7x6dMH4+fJYo/9s+1UaEg2pt1ny5eQy5ivVmHyYKyLqkh5fxuR4o7UpJMo1Igrco/V343PUWL5+vSsWnPXy3V8wyEvtxpmg7ETRZnKmBNW5j1m6dhPTq1Tk9G+vdShpN7o9JdExFVkwDnxIla6Xwbnr0h7KKNlUk3X87l0sOeiMq4hMlIqHZ02qBfPkhVVlLN58Uqzb+JuMjXl4zq/uSVber1eLtf3afDx7evgr//x6q9StZC5ZYmw+YtYvezvb+Ei+KhVcLmZ/ThfS84P/9ksOXOyI8Q1LJup1D7rNX0cPxjCqTrwcI4M/bta5/oyMVoJCBWEas+E6of/LPv7RJz+jjldS1Zqi9b/rHOSiTuOMcyF4HCRdIPInwmW0Soib/4DOeecTkbcE95wsteS/qjCLNT5X3yWSodKaTlZlAQtDzNWHouvtN7scwm9DBAtiNYeidZfbHt1EesE+EvF7WX/WrzO6yxvFmNBh86cOIG8vtJjjNfSGV4Y8Src6qqItRGGUEGo9kuoXjX11kOEpJQwq4U3Kc/m/EbGSGrxl9fyhE5bHIv76in5TJ/K9c3Uj79+UStElvJv68WDTAbajTTJC6O2uO8unr9CLvVx5lIjcRqJ00icRuI0EqeROL3LxGkkSiNRGonSTyU0LXmwQxluVtw5evjONPCwTsu+j9yuYkDtmyRfwKtuzvtyMm6kMWLFvKHVZwaQTKb3iFPiKgyXccbfjaB2G3IqWJOH89/N3bJKnOizeXjHWF87H8Nbol3bhNdD8Xoo3jGG7sE7xtBEx6eJ8I4xHhUceFTHo4J4x7hro/GKoCWRx/uK4Lb8jbcCzW7jHeNtVxDvGMPlgMuBd4zhgRyTB4J3jA/WN8E7xnjHGO8Y4x1j3zvG7ZFVZsJaJNWOm75I/8ydKWywA15WzzVJhHCvKxa5LkuJC//pp3cl6FqjS1mUzkwyepLPqzHMknG2fi4kujZeolgoeD3nUxdu4n1EC3sl1LMhWRTnoe3Z1F4Y4e2P0moKvAjTSGaOJ7+/XF9fOz3o917EVxIyXg73Baw0c/PqysxFnhqDaxv/1vr6SD2XDk9/4OkPPP2Bpz/27OmPxntp+4Bfefeo8Y3Bzgfuhl6PdWvBEqzRquMsTIe0M6L9ervefPbDqx/+Mnv1b7Pvv69ca1sj6MJrSL/FQNOIYXP1UonJ9VevPM7S6P88yrH0ARdRvmqTikuiYeEnNOlTxeruTboiVuhxdbE+jrNb/CHO0ysALp0L1clv6g56J5cAFtyifjrzbOynYje7EBwqsnRHA1snM+mBkFO1C2SMIhlIktYopJutl1IFH/XhUoHX9soWNicj67nwPBls92DOFYUWOFmFX2c0lFkdMmPH2vQ+f/+qpdt34ddola8CNTFLoEzOv8Jk5kkcvPj+5V9eSXBFEsa/06Rn603V1fZwKH8w8K0iJ3qt+iP2u1kntOUrW7UwomKW3dJ4b9fLRd2BtVaidSGiWC5EUgGdFJR1zEwteKfy8PajfkoVNTMvOqZxbzII7Sz/qABITRa7FVeJoWgOnz+b6tLZy67esn85RGnp75TG2gJiIbOo9amtBFqM9SvfuF5l8Shr0Z16mNXF6O6rmvO+k2tm+6e08kruli6z8ZQVGevpNvmAzgjEkhay6mO3PsHy8WoFWvGEMrWab1xfS//3AEmiT8Y/K+2enPVL2uZd3oJk73sZES8266jMvXWp3YpwsdcCI9+V0LOQSxfPl/lCkJqQz6vFMZGQMmR8BDNzzbMzciGSNTF7v9OG2uLWo0Ms7typ+uC4itR4R9BaBeMTSmFh0UuVEtYcqecaFHNl/S6H+ai9yy4aOu97Sj3XX7yVDl+T5WVFVdcNzOJyLKZL55R8Vjm+qYWo3SA3fKqwb+Xjyy8+601U5yhJSB+jlI8qvcGQX7gQi/LZ/RqXdS19D53Gb0XKld/tjhZX/jXGUiF5ntiA/VQfyBe+h+4pKXppoViW1csoZpedAPFZQwDTmtdHQ4V9iVgTqhExb5VIKPXw99C1Baw8hW7p7Qbl020pabq11WotDjHUcI5KQbJiHzuMbCOmc+wxHSf4spvbd0R4pu3JIcKDCM/hRXjKaMz2WgzBHgR7EOxxxYujM1tLFkI++yQ2CPkg5IOQD0I+CPlsG/LpAOfLooTfZOKo1LC7YTTnnCeWC0eH6yxgNxOWznpRfM3p/xV/b1QavHpk1yTYWrug3gDmdQgjsw6m1oRcqpIK55DxyGSZXfp0s05V7R7GCpx8d/f9d2tJhn8qMqQKjkilJpDm/OQ2yzY/fvddGD+8vCUiYvFyuQxX4cs0C+e/c3kbyd2pmOd07nuQH74xeuzXL/w3zkf7WD5H/Pfy8WT9nHBpr3ROcfkLNyO4/H1zInD5d51JK1PpitxZ9V9OtuyrMju2lv9aEFOHD4u4SduscZHVv/KD5T6UeWcNVrwxDGdibZaWdUNB9WDOr4bkFzvU0hSJ4GhDQ3zARADseVYO7q9stW8d7KoL5FkU50j0rbJIJWHl3Lpuqr1epa9Z/rb0GRvdqLJhxfKVq8acWqgRezFr0v1KKpzrtWQVLZo/sUwElywTwaU2UAYNxQKlVuLu+7rXrpNh7XfRnaRzm7I0c7oqavo5Dir/C8k0SBNp/OvUAAgiNl8G58DmihVTch3OLdv5mUFGy3Uik9jrxK+4dM2S3QmGRKhRSV2vjjmsYFahPEtrbuh6sbxyAi/epOv8ULMJi8t3m2Wooov67KEU3K9qpZWK4x34Xto3o+ZoU1mVcbs//rgKU/FLsvz2jX/ND5k+qAfUZcD8ijf31y/FoURK8O/i4aR4c+flJx3RYcAG253q6+rM2OqL8/lcbLLWtl8slf3h/eUnanylX3Bn1Ui/TcJ7ngv9/48ncoqW/N7zP8swvsnla0sniib/7/8BacwhIA== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Create a new OpenAI response. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/create-a-new-prompt.api.mdx b/versioned_docs/version-v0.2.23/api/create-a-new-prompt.api.mdx new file mode 100644 index 0000000..f3b1918 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/create-a-new-prompt.api.mdx @@ -0,0 +1,68 @@ +--- +id: create-a-new-prompt +title: "Create a new prompt." +description: "Create a new prompt." +sidebar_label: "Create a new prompt." +hide_title: true +hide_table_of_contents: true +api: eJztWF1vEzkU/SvWvABS0hYEAqrVSoVltUigVm1BWpUKOTM3GYPHHmxP0ijKf+dc29NMWwJlhbRCah/adGLfr3POvWOvCke+tcaTL/ZXxaO9Pf5TkS+daoOyptgvTmsSpSMZqBJHzjZtENhkO1fSTjEqSmsCmcD7ZNtqVUret/vJ8+ZV4cuaGsmfwrIlmLOTT1QGbGydbckFlVy30fJgnQ9OmRnW3YzGL32gRqQtItBFEAsVajGXTsmJJtFqWVJtdUXO74j3+bEX0pGwRi+F79rWOs5oUZMRnYcrEWD6uC+HODh6vVOsR8UcNlTKJUemkPCM3I3Q3qeV4n5eIHyQyA+WZRAPR0IZlLFBreAWy7yc0wP2kPL4qKofZ//OqC8dCVXBipoq+Jha18jAJqUX91o29MfjZ+NKzVQY19LXf96LWfQ1GPiQzsklXCgU09/0vb7u/I3yQdhpX/fLahvZoF6hRpalNGJCqCfiUSZW9BKlBqgEYrPKf6xoKjs9xHtirSZpYsr5u6nUnq5H8SKtg/kqcg3lBYbw5OBOeZHxEspH99nY5WPUK63LhFsjHllViq1LfTTgZPbu6EunHAGcs0suDDEb1vZKauejIqigObckmxtoHvTF6fWEDy0+M7jMGhDIwrM4bMkcvBYvsRYJR4JfLSrX+o2WjRQnQZafwds15/V4m5w5JwKYC3BGmbnUCox0opGa6UTVr5M1JBA6fwvx/HN6eiTSalHaKvIkl+9HonjlHIPKi0dctBrKhsKbRrol85VpQHHNolZlzczgpAGaCZEPYFP6PvqJvA9S6dt6TqvZtbaGdV93jTRjtMwq6mOw50o4UQoGOZvyFlneP2wTRx+IA/Hu+E1OZqi4YIEsdtOcRAPmIM3UHtiznNguROe+pRLNoxS2LDvnCO6vxfUzmsgI92hdFm9A/1imLcXr54+YgtIxBm684th2sxqNemq1tgsvjv9+KZ4+23saWzJdSPCehvwC1Tf+XsgKjTxSfBPPLZmflfPo+ZZBqBVUIdBYBcsUJbfYbJa9YbYKIszUHGNFNrYzsWUG1dCdqO5E9ZuJ6tHzjb9TMP0tMz0ryw+l9a/toAlUiC5KoorLxmLj0aRVo8KOOMLQRkAB7JEzCZHw3HL9pHqybVJ5chi6AuVkKRFPQ2DTGbpAvfmth3IZ7qR1J63fSFpPhvPqNTMbkYuTxPaN/yyvg5uU74tR7YjDDkwg2cSpNCEMHmPj6aDq5TV42772Bvodw3eiuhPV7yWqvW95u4WEokqYanLmOZt0XPMc/dVoX8ZrEGBsaJHPYHwF0kqHIzA0zLvPU2UwIV/Yasnh/Q83JMOrkez+e1ckv/qG4L9dDez8HOdyPQYUS/Ak9Pr3/wjsZltwHeEB0Kot/i1aG88IOFfX+G93/nC3zeCPivTuwZiuis4xieoQ2v3dXbwFjWtspGqs+dQ99vHUXdqmWJ/zPtBKhWXc+Fffes/O+TsG/XhDj1cbEvfwbmAdIHLWPz2P7WJqIxg57cHJX5xkWUeKFYPbq2L+8BtM4S7or/QDOexQQ8sRRAQ5VzDywYhrP2i5FLEnU7UWDd6jo8bXQOVAQZckXvG9Bnd/NwX9MjOkow+GZRovO4K9aXzCRzaNjubkrI/qLeaDTtRlENFsuSZMuO1CvZL+QJjb1me6s4x2QVFl2FukwipT5izVtd10DCYGP1+tJnjhfef0es2PAblbpvbQ4xo5gQ0YEZiFTJbPtORgUljjU/bOy3UXxXi9d6xH/Y6DsqR4v7R97fmA9EeHJ6dYPMn9qeFJu184ucBD/r1f4IONZYqqj89WhZZm1gEAfJ9s8s9XtOrPKw== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Create a new prompt. + + + + + + + + + + + + +'"},"variables":{"type":"array","items":{"type":"string"},"description":"List of prompt variable names that can be used in the prompt template"},"is_default":{"type":"boolean","default":false,"description":"Boolean indicating whether this version is the default version for this prompt"}},"additionalProperties":false,"required":["version","prompt_id","variables","is_default"],"title":"Prompt","description":"A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/create-a-new-session-for-an-agent.api.mdx b/versioned_docs/version-v0.2.23/api/create-a-new-session-for-an-agent.api.mdx new file mode 100644 index 0000000..51ffa61 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/create-a-new-session-for-an-agent.api.mdx @@ -0,0 +1,68 @@ +--- +id: create-a-new-session-for-an-agent +title: "Create a new session for an agent." +description: "Create a new session for an agent." +sidebar_label: "Create a new session for an agent." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETi2Q2GmwRbe5uWkXG6BFAsc5LNJgQUtjiy1FqiTlxDD83/uGlGK5TrbuopcA8SGRJXK+3ntDjVeZI19b48lnJ6vs+OiI/xXkc6fqoKzJTrKREaM5mXBJ3uPOqSMZaNxuG2QHWW5NwHPeKetaq1zyzuEXz9tXmc9LqiRfhWVNMGinXygP2Fg7W5MLKjn3yf6/quit9cEpM8fa7ZiujPrWkFAF/KqZIidm1olQkshjeIVorWXr9UEmi0LxRqkveh5nUns6QAG+NcoRnF73Q7g5yIIKmmN4PPuduLoHwlFonEEctyWZFBTyEFIYuhWSDXYRDhAiB/nHQ7WflGwKqfogbqUXyiykVoVAspXUyLlC4L8PgSBD43vrFMzOye2k+fdkciHSapHbgrL1fbV+htwH5xgpXnyAcvjSOpSiqSrplsLOIoYU19yWKi+FSkk7JVEyBlma9nn0s2b7QSq9r+e0ml1ra5CbKJtKmkMAVMipJtHbsxUOe1IGOZt8jyxfnNeJcC/FSFyNP7bJ5Ah+SqLxIEawTBKnaEGisg5kNoxnRE7IqW1CdO5rysHwXNg8b5wjuP8hrl8ieEK4Q+u+eD22xzI9UryuWYiZs1WMYXRxNhBj28xLvQQ8WttbL8Z/nYo3fx69GXDR6E5WdWJGxy9QfePvnSzEOFF8E8+ezG+Vc/z2YeXkWrHQSmz3fBGsxWaz7AyzVRBhrhbQqKxsgzVcW1XRs6ieRfXERHX8duNvAqZ/Yqa3yvJ9af1jG2gCFaK7nKjgsrHYcK4JrSoVBuJCk0RAAeyRcwmRaDx03Un1+rGTypNbAHyUk6VEKA8TqzF0h3rzoUxtGZ6l9SytJySt1/3z6oyZjcjFZWL7xn8rr9Eu5btiFANx3oAJJKt4Kk0JB4+x8RW26ORV0Ew2Ojz4Iv6o4WdRPYvqaYnq6CFve0goqoSpJuees4nTmefgt4NNk1o7cbWzVsedOIDx6FpLJyuCoNnUKjP4gq3xMQ+BjDa+1zKUO8XgA+/sfQdDGukAZ5o/E2obp+xrA0FwDSDZEWRLo/X6Ji3Guf3OFkte8ZtH7JTnz/jLGfLKLscun/ssB/9vuo7ee2RLSPWn7G4eiEBv1w03AFhpC8bFxpkhwnOSDRevhhEGP1x1CK6H3e8AqHfs1wnoxjHNyhDqk+EQ70mHJUxRcai1rOQhKJp/HeS2yhgKTyCeCsu48X3XnK9v+BkDMN5A9aFH861S32PLjJrZWPs2+4/sUlyyS3HZ6jyii5A53oTF4tUD8HBb9FsNQvZbVt9yKGUQIMNCwchnI374oAdTnH3IFLVFx/eQSXwvVA7N3SXNF/wTBh8HbiZz8smodPTZsG6tiw1t1/iUZziNFucASxvVJxwY2sfewCii+3JN2mLtJd2tYvQUst/ulvmB7sKw1njF5UgiK1Ytn65TzROjcHHS6wodqcAAJg4vXq2meGW+cnq95tughGPG4DKeW1PGGvwplOfr4l4gj6bxYtzy/qX45VbzYHbtTbCdmSV1w99w+ZWW/Z63vsHaEichjnyOOD0+TXEdTtjIZvtOM1ofdDtGeU51+M+1Nz0tX5xfTrB42ja8il8oTjInb7lz4m8M1cY6xVYW760yLc28Qex4nmzy5zsBj0lx +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Create a new session for an agent. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/create-a-new-turn-for-an-agent.api.mdx b/versioned_docs/version-v0.2.23/api/create-a-new-turn-for-an-agent.api.mdx new file mode 100644 index 0000000..893862c --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/create-a-new-turn-for-an-agent.api.mdx @@ -0,0 +1,68 @@ +--- +id: create-a-new-turn-for-an-agent +title: "Create a new turn for an agent." +description: "Create a new turn for an agent." +sidebar_label: "Create a new turn for an agent." +hide_title: true +hide_table_of_contents: true +api: eJztfWuP20iW5V8h9KW7gbSqZzC7s+PdWSDb5ar1wp4yMrM9WFQXVCExMsU2RWr4yCzByP++cW88GHyJklPVpqjTKLhTfATjeeLGuTdOfJllMt+mSS7z2esvs3/+85/p/0KZr7JoW0RpMns9e3cf5EUmxebffxBxLq+CTBZlluSBCO7U/wfp8u9yVcyD6rm7rPQfS4Lb27eBfJRJYZ4I0vvg+kH9pgRuTA5u+dabdZl8ns+uZqs0KdQTlB+x3cbRSlB+vvt7Tpn6MstXa7kR9Fex20qVTZ0N9eI2S7cyKyJdJMrDIgq9B1UWouRBPVgv5l+T6L9KGUSh+mh0H8ksuE+zoFjLgJIInqJiHSWqzKqqcnrj+Wpm/nxh8qqgjzLLuXh+6lGyLYvFRl0QD6Ys+gsiy8ROfSAq5Iavp4n86X72+uehusjSWHblVOUgVxU9K3OZccbvRRnXLvgl+VDmRbCUwd/49t9mQZHaYu1UgSLV4tQ36GZgck/F8dqznt6drgO6Sf2CqsS8dhU8raPVOlipLhQlq7gMVWPI3wrVpcIgVQ/SB8JIqCy2qsAU7vnq8OrRd/urJ9pQUfz6qa74Bfo+ol+bKBEFNbFKzZbLFpNabh5cx09il6t65GT+NuNG5xQHe3WZxe2KvA7+evPefooTCtT3Q1EIvqF6L91Q3W4juKbpzmt+7rsv9Lnn/7kUufzv/3L1he48z4P/SAtV4Wv1NL2/Sss4DNbiUQaxTB6KdRCrQhY5Ddah3EbDA4S6AX1G3w62aaR6lPpD9S7KtsKptMxWqiupWhJhGNFrIv7ofeie4Ukh2n+VUSbVmPyZv/yLyl5UUM+fqfSpkql4PQ1NzfM2WaUhXXk90xXSyqu+HEh6UIamrrmiVde3PW8onw2U5TR45DRSNxWimlINAy7BMTXAZbT9yqsK/t4bXeB3qju2ynhtiuX3WR5OLxtENH5rY8hdeNEQolT0COL0hnsbwYiFpK+pUP6MV5+U4P7qLLxPmtpU74d+SSnfpjp3/yE2lK753EbNgdwlvxwKEUAzoBnQbADNmmgBUHs5qNGHXbuoG1ks1SgL/XS8aeSrjVngG/AN+AZrDdYa0AxodhFoBmvt21hrz7/se8aRe791kHt//Gmri/Sn4I6owftIqoGs/ihz6ryUmLoZ74LlLngfi40Ibgux+kwjbyvyPLi5/jEwic/9FDZiRxxkJjfpIyfELXP98Z2DmrIoMzkHNQh4BjzD2Dw7XIaxCTQDmsHYnBSogRoEvgHfxoFvADZYa0AzoNk00AzW2jipwWPyzuGJVaCgP0BymX0woYTtrJtoweA+SzfcChx6yMGaK0IJP8ByfsDMNxQmWaRpXO8k9kJ3mCTd7g2TpJuBDcJlKlXE8UsDVilNSkd/x6ZOtKt6ZDAY0z1vu4OrWFNQMKqY1TCrwUY/r+kMNjrQDGgGG31SoAZGFfgGfBsHvgHYYK0BzYBm00AzWGvTYVQNpdjJrd6laWy34A9zrJncqgEqzYjVw1W1K7WW4TKj5DFdWa71l+dmUu+jnJ+2+9o17ESJKpAo1CBgxpJ23PMG+0JuT7Lz/WAVAAKnd9/bzkevzW1GvuJ9es28LzJVuoXo7KoaqdUVhSzyVRFt2i1A6dINl2xgkpxrOnezjeWJP+ASrWqgb7DLpNxQf4uSe5kpOOMBozrDQv4mVyV/QaWwpsjdBXVGmrXlJs12i0yqNOSjiP0ueau+daeHXSOTHjBwHonkV/+RlIRpLG8W9zLjzeTuqirURuFuvHDs+yDYDXkGRJ6r7i2SOuLVr3b7CNwzPY4CVkWg3P4hr3sLjlFVaL4PJh/WFKwprA3PzYzC2hBoBjTD2nBSoAYmH/gGfBsHvgHYYK0BzYBm00AzWGvjZPKJVE23i0wKo2nbR6vKJFyk9wtmxa/sr43j6tOy4NvpZ5nkVLJ6/m84/eBpvatYwIC+vFU99UEmMhM0aueBlmpQ3TmTr4NXwa+36hn98tzLwa+vgzuXzH2URPm6lg5/hNjLrIoknncnZ4owmOKyLOxPdUMEW6EaWVRh08GrV0GZl6QjcWXdEEQzzzldLUQrdjYi2sRHk4YstWmUlLKtw0tyv+5ZjzNtlqRW97WCZGroq7vc++muKkX4IAsm05kbp+zsdW7sH669ceI2/YTn1o6VQat/LTPVLxe5FNlqrS4/pfF9JjYLEW/XpK27XadFquqfsSCUC1bt2GZS/euPo7+UUaxqk/xJvgllc0WuIJE9lBvVN7pXLB733Cx69xg/KA3K7YPM/Euq2Mv6laXKtBRJ/aE4Pm7Fdepv//I8qqpQ2eH/vGZc/L0TuY5C5co5WfVav6c0XJVvBDdLj1fRDf18HrwVq7W3FSJiwXKTghUtf9EGlTp+e/l8oz1nKslhhyolJhTmGeC0Ppo/OofQn7z9G31bWo6aBY0rtHJq+s69lk+sNocZ1xn559pFIo1u80Cfc+6AlRUctRfnqG1loL6zyruFqRNTJ6bO006dd/6GwZzW7nq8ycpSrZ0RcuIhd8pdjaMcu80CEuXlpi0qErM1TyLnyKnPqmXrYR0I0QBNBpoMpP958WMg/YFmQDOQ/pMCNYRoAN+Ab+PANwAbrDWgGdBsGmgGa22UIRpdo2lAHY12lclCNAbD8RR0k0Pv5JabBHSbt9ZGaHu67NiZ6R3Vce1yGdjCBGJJ0QyOtaw2ob2csR7YldpqBntDV33XvtNekttR2fVmMz7j03pSPVdVi0hvFPWtdXF1u1Z1Hp0fDO5VuFcPd6/Wv1fNDf51VbDHKI2FTr9rL6m7XQ0cnYCOtBq0Vtz7i1g+yg4761ZdzqJiF/B9W7oqV1f7ajsl/43Q4XFZltYcNJ9sEu/5w6qoFBHm4ucGu7KHiyacg6wnjsDYWTuKY8wqhKxyfb7TQdccoA9hJ7sp38pVdB+tvH5BdhW74oJQLsuHB3qMguwKGUuVQrY7Dl2bHcarSH+IiHtZ7D5V9X06APe/wj2d/Kc9+OwNBYAzwPlwcG59wUfo1k2CadVz0mwRLlVL54c2de7glN/mRRWth9jdb5KX9pw1BniuzkiZKtwjzB24o0EJgBIAwXl+XAAITqAZ0AwE56RADe5o4BvwbRz4BmCDtQY0A5pNA81grZ2NO9qSdpbFCyv/TJPsO7F7s85EdrCFXmE+MJN5Y4nMHhZd852B4zt7qfQDJojallZvlnDk73DvBh1/cXQ8NIMtaEEzGFYYrDCsKadifmFNCTQDmmFNOSlQgwcA+AZ8Gwe+AdhgrQHNgGbTQDNYa6P0ADxDMxiawV6zQ/iwtcXl26v9Qfhwdg7Ch9AMPhPN4KZbEJ5beG4hIoy5FHPpaOZSp68AEWGICIM3A292qbwZCDN4AYBmQLNpoBm8AIjZAL4B36aKbwA2WGtAM6DZNNAM1tooYza6RhNEhCEiPFoR4ZrvDs5WOFshKQxJ4VNMDpAUPrWkcHusA6+B11AZBmcAzgCcARhQMKBAM6DZxaAZGFD4q4FvwLep4huADdYa0AxoNg00g7V2Nv5qS9pdlMowV2b97Z+yUKqMK7jSe4dVT13JPGf3hUolt9uewiAsM70/N8o5PepIaVlsy6LDTQTpWCwUMLViasVCYYxzKhYKQDOgGRYKkwI10LrAN+DbOPANwAZrDWgGNJsGmsFaGyWt+wzpWEjHes0OubvWVoZvr/EGubvZOcjdXbh0rPNkiaJQpXZD7CuRBTJ1sNFho4NxOFfjHIwD0AxoBsZhUqAG/xDwDfg2DnwDsMFaA5oBzaaBZrDWvpF/aHAeGdUoa5ebvppJe1wUnzGiSkpqQqYy5h0nZXQEplespZbQiTayV86lndiHdx/e1jpSLbmjqGhH7FZZ8Crh2qXbrovE+yrVRFsYqP6CJzj1QxTLnLBgI8NImHQUSJj242QaUf8vl+xRF1Uim23wtJaJ0xMKlvKBqf9TqPZ4Rez7nHX2XQXRfaXp89X7VWjjR2p+RIm3syOf6T0s+ay156NWmf6o187WZpennSWxyb3eocLeMLGqvIbq2jW1WbQKbnc5gRkXiEDiO/mobrxStSjFhqo1Vy296VQNqwMBv9c2D94lYfQYhaUacfyENQiqrlc5D8xHhyb2rdjFqQjb33qrs241wcxzvv9CZ4Fn4o6dJftKNijexBuQuKFMSy5sb7G/VaoPqqwsykg9wj7MP7yH+bd4EhFh5IJ7CTW7ncRqH/J09OqXa0PJII8u/VJSTWRym2aFPvTnHyFO1Z0j3kalM2R3QR2lF9Z/aBOnTJ2dPVJ2N9WZiuANKKT6ZT4OmbyuXd9CZ1vAn1toxBLmWBVU2it3Sx3uoxmQzWyb65y/SnRNd8OcsakNArQXcNDoGOOI9NJqDMraHYzLrnEZylgPyubkscok2SuVvUbO5URNl3azJKfnF/lULOJL1wpcpJcsEmg57ZW/+D0WDd9zvQ8sF7gkZhDTBK276RHD9fenOJqVvYfbOH7d3WyJ+sJ7uStUKU+3Ou5pkqSxMP7dG8UFgNUHQe3qV44Em4YZDi7JIyOuEGyGYDMEm3GwWWsif6tmWJnpxd8ra4B4oWYVvcfsbET25FZkOS/qzUM0cvga2UVFme+3tXiFyqta3+C5F1HMl/NytZIylGHb6HhTZhlBGn1LT+n0tdpRpTo7XzPpeXhVK0tHNfZOhq7WDgXfr/IXjIPivCCzZTweGpgv7capmQUwZWDKwJSBKQNT5h9tyvheQuMa1Imflme0FNAQ3/jRtMfBlKPrMNNkHb3HG6xj7Q5Yxy7WkRMOZaEG9nEBiDgO5eKOQ/Ez4xn67io5l2hjlDtqbdhshVQfwuEQDvfV69oLDof71iTRpOLgENwLNAOaIbh3UqCGrVjAN+DbOPANwAZrDWgGNJsGmsFag1QfpPog1YeQA4QcTCTkoIlxFy7Vd9QsOHAUV8MnVpvDjOus+8gtivaw23b7DtsaXlnBUXtxjtpWBhoRf9UtTJ2YOjF1nnbqvPPD3HItuMCRPJWl6mKufochd3hEUDWVc1BENSvm9Mhox26zgER51cILma15EhTo9qgWCmEjrAMhGqDJQJOB9D8vfgykP9AMaAbSf1KghhAN4BvwbRz4BmCDtQY0A5pNA81grY0yRKNrNDne0X7tPks3js3jXWWTlCwUS4pmcKxltQnt5Yy17SoNvvqm2qfWjGgxTcBVbwI+iD1dVWeX9ZDcjsquN5vxGZ/Wk+q5qlpEeqOob62Lq9u1qvPo/GBwr8K9erh7tf49b8+5d10V7DFKY6HT79pL6m5XA0cnoCOtBq0V9/4ilo+yw866VZezqNgFfN+WrsrV1b7aTsl/I3R4XJalNQfNJ5vEe/6wKipFhLn4ucGu7OGiCecg64kjMHbWjuIYswohq1yf73TQNQdEySouyYgLnGRz1S/IrmJXXBDKZfnwQI9RkF0hY6lSyHbHoWuzw3gV6Q8RcS+L3aeqvk8H4P5XuKeT/7QHn72hAHAGOB8Ozq0v+AjdukkwrXpOmi3CpWrpTlGcrqZ2ejb6bV5U0XqI3f0meWPQ/qYNWq22pkwV7hHmDtzRoARACYDgPD8uAAQn0AxoBoJzUqAGdzTwDfg2DnwDsMFaA5oBzaaBZrDWzsYdbUk7y+KFlX+mSfad2L1ZZyI72EKvMB+YybyxRGYPi675zsDxnb1U+gETRG1LqzdLOPJ3uHeDjr84Oh6awRa0oBkMKwxWGNaUUzG/sKYEmgHNsKacFKjBAwB8A76NA98AbLDWgGZAs2mgGay1UXoAnqEZDM1gr9khfNja4vLt1f4gfDg7B+FDaAafiWZw0y0Izy08txARxlyKuXQ0c6nTV4CIMESEwZuBN7tU3gyEGbwAQDOg2TTQDF4AxGwA34BvU8U3ABusNaAZ0GwaaAZrbZQxG12jCSLCEBEerYhwzXcHZyucrZAUhqTwKSYHSAqfWlK4PdaB18BrqAyDMwBnAM4ADCgYUKAZ0Oxi0AwMKPzVwDfg21TxDcAGaw1oBjSbBprBWjsbf7Ul7S5KZfi55cAwm3TVAJWFiGJHeJrNTiEndVz5VWUmhS2sX/BGdZgv+sW9pszeqbxaVzYV2Gbxo9jFqQhb5TfX2anic9vkbuG85FwVrh4qZ/chRxh6pdnDjXOBmLS3pbMEu/2tUn1QH2afN3UQ+zD/8B7m3+JJRITviyjZltwhHDT473oAUb/cRatzQYKlJLjM5DalLsdYcKj3ZM8WNKpV0orgrdmqYvOjvUy1TmOztK9n0N+3VOBDugVnkKtnoj3Ce7zRKWp3ju0XJ/S9Hd17VC3pP1+YfE3QxEudK9I6uPdu4Tx0wT+ki03+71oDuQt+SSo1bLrdI4QtjJ6Lcc8/HymBrV+7Cp7W0WodrERinNZST8DkkE7Vg/SBMKJNn/BeYQWBFQT4kPNaOoAPAZoBzcCHTArU4L0CvgHfxoFvADZYa0AzoNk00AzW2ii9V5bc+62D3PO23dwRNXjPOyzUH2VOnZcSS0gQOljugvex2IjgthCrzzTytiLPg5vrH60nbO6nQELRSxqZm/SRE+KWuf74zkFNWZSZnIMaBDwDnmFsnh0uw9gEmgHNYGxOCtRADQLfgG/jwDcAG6w1oBnQbBpoBmttnNTgMXlvnNvkD5BcZsNnNLn4eA497D2FaXDmGwqTZL24WiexF7rDJOl2b5hkQ5Tt6h9z4sbeYMwBpTwwqpjVMKvBRj+36Qw2OtAMaAYbfVKgBkYV+AZ8Gwe+AdhgrQHNgGbTQDNYa9NhVN3pDR3cqn9iwzDHmsmtGqDSjFg9XFW7Umt1nefQFiR/H+X8tN3XrmEnSlSBBMl5MGPJW/uftQzBSXa+Q4H78hS4/cx4M7m7qgq1Ubhbne4xDHZDngGR56p7i6SOePWr3T4C90yPo4BVESi3f8jr3oJjVBWa74PJhzUFawprw3Mzo7A2BJoBzbA2nBSogckHvgHfxoFvADZYa0AzoNk00AzW2jiZfCJV0+0ikyKnD/XTqjIJF+n9glnxK/tr47j6tCz4dvpZJqxWXc//DacfPK13FQsY0Je3qqc+yERmgkbtPNBSDao7Z/J18Cr49VY9o1+eezn49XVw55K5j5IoX9fS0ercCgiyKpJ43p2cKcJgisuysD/VDRFshWpkUYVNB69eBWVeko7ElXVD8Hm2nK4WohU7GxFt4qNJQ5baNEpK2dbhJblf96zHmTZLUqv7WkEyNfTpJFnq/XRXlSJ8kAWT6d4J03ucG/uHa2+cuE2fD+nuWhm0+tcyU/1ykUuRrdbq8lMa32disxDxdk3autt1WqSq/hkLQrlg1Y5tJtW//jj6SxnFqjbv9GnqrVyRK0hkD+VG9Y3uFYvHPTeLfvAJu+00KLcP9RN122fsdh7EWz+rd3jFdepv//I8qqpQ2eH/vGZc/L0TuU5wtHzVUxquyjfmoOtur6J3JnzwVqzW3lYI8qMENoVAV+eRByc0NqjU8dvL5xun7z/sUKXEhMI8A5zWR/NH5xD6k7d/o29Ly1Gz4MDpDw2fWG0OM66z7lMeSKPbPNB/ejIctXDUNh21rQzUd1Z5tzB1YurE1HnaqfPO3zDI54abo3UqS9WdPvI7DLlT7moc5dhtFpAoLzdtUZGYrXkSOUdOfdZnm3hhHQjRAE0Gmgyk/3nxYyD9gWZAM5D+kwI1hGgA34Bv48A3ABusNaAZ0GwaaAZrbZQhGl2jaUAdjXaVyUI0BsPxFHSTQ+/klpsEdJu31kZoe7rs2JnpHdVx7XIZ2MIEYknRDI61rDahvZyxHtiV2moGe0NXfde+016S21HZ9WYzPuPTelI9V1WLSG8U9a11cXW7VnUenR8M7lW4Vw93r9a/V80N/nVVsMcojYVOv2svqbtdDRydgI60GrRW3PuLWD7KDjvrVl3OomIX8H1buipXV/tqOyX/jdDhcVmW1hw0n2wS7/nDqqgUEebi5wa7soeLJpyDrCeOwNhZO4pjzCqErHJ9vtNB1xygD2EnuynfylV0H628fkF2FbviglAuy4cHeoyC7AoZS5VCtjsOXZsdxqtIf4iIe1nsPlX1fToA97/CPZ38pz347A0FgDPA+XBwbn3BR+jWTYJp1XPSbBEuVUvnhzZ17uCU3+ZFFa2H2N1vkpf2nDUGeK7OSJkq3CPMHbijQQmAEgDBeX5cAAhOoBnQDATnpEAN7mjgG/BtHPgGYIO1BjQDmk0DzWCtnY072pJ2lsULK/9Mk+w7sXuzzkR2sIVeYT4wk3ljicweFl3znYHjO3up9AMmiNqWVm+WcOTvcO8GHX9xdDw0gy1oQTMYVhisMKwpp2J+YU0JNAOaYU05KVCDBwD4BnwbB74B2GCtAc2AZtNAM1hro/QAPEMzGJrBXrND+LC1xeXbq/1B+HB2DsKH0Aw+E83gplsQnlt4biEijLkUc+lo5lKnrwARYYgIgzcDb3apvBkIM3gBgGZAs2mgGbwAiNkAvgHfpopvADZYa0AzoNk00AzW2ihjNrpGE0SEISI8WhHhmu8OzlY4WyEpDEnhU0wOkBQ+taRwe6wDr4HXUBkGZwDOAJwBGFAwoEAzoNnFoBkYUPirgW/At6niG4AN1hrQDGg2DTSDtXY2/mpL2l2UyjBXZv3tn7JQqowruNJ7h1VPXck8Z/eFSiW3257CICwzvT83yjk96khpWWzLosNNBOlYLBQwtWJqxUJhjHMqFgpAM6AZFgqTAjXQusA34Ns48A3ABmsNaAY0mwaawVobJa37DOlYSMd6zQ65u9ZWhm+v8Qa5u9k5yN1duHSs82SJolCldkPsK5EFMnWw0WGjg3E4V+McjAPQDGgGxmFSoAb/EPAN+DYOfAOwwVoDmgHNpoFmsNa+kX9ocB4Z1Shrl5u+mkl7XBSfMaJKSmpCpjLmHSdldASmV6ylltCJNrJXzqWd2Id3H97WOlItuaOoaEfsVlnwKuHapduui8T7KtVEWxio/oInOPVDFMucsGAjw0iYdBRImPbjZBpR/y+X7FEXVSKbbfC0lonTEwqW8oGp/1Oo9nhF7PucdfZdBdF9penz1ftVaONHan5EibezI5/pPSz5rLXno1aZ/qjXztZml6edJbHJvd6hwt4wsaq8huraNbVZtApudzmB2XEFko/qZac2SNnw+yClTHmzaon0t/E3yI9iF6cibGXbXNdn7lDOV85BEfDXci6L66+VnuL8AGPXy+8eHzZvIeKqNm2xsO1tf6tUH9SHc1Nq9zD/8B7m3+JJRIRyC25nqiE3DXXcrs1KPfdrw8OgCRdNDQqC00xu06zQB/lwqwxPmIcKe+05HYlag3oVu35MB2cEqPr6C5Kvuby91BuDZ48n6VBaYWjnFDnpa83kLnTvl6LbPVulhPH4mwF+7CYp/dqVQqpotQ5WPMJJEU/qaZ7CBlL1oIHrGfZQYZ2CdQpYl3NboIB1AZoBzcC6TArU4CMDvgHfxoFvADZYa0AzoNk00AzW2lj3UK2M7FQLF3z/A1GD93x8g/qjzKnzJtpVFe+C5S54H4uNCG4LsfpMI28r8jy4uf7RymzN/RRoK9GSRuYmfeSEuGWuP75zUFMWZSbnoAYBz4BnGJtnh8swNoFmQDMYm5MCNVCDwDfg2zjwDcAGaw1oBjSbBprBWhsnNXhM3hvKHv4AyWU2rOLhxPc59LBfp+OlYZJ8GH2tk9gL3WGSdLs3TLJx4vtVv6TQEQGrTm+Fv1NJl/CpwYPBmO552x1q57mDUcWshlkNNvrZTWew0YFmQDPY6JMCNTCqwDfg2zjwDcAGaw1oBjSbBprBWpsOo+rUnTu4VVJjtjoGwxxrJrdqgMrEicyrX6pdqbUMlxklj+nKcq0kVF1PyopE233tGnaiRBWIheVrR4NquYjXJ7BfD1UBIHB6973tfE43xJ7CeuT79Nr8NFohlBt1wyUbmCTnJxIKaX/AJVrVQN9gt/oSUWJ0aGjAkJC4PviVvqBSWFPkLivu06zN580u3HmzfpekI2rv9LDrloRweWyfT+vN4l5mvJncXVWFYqHvhWPfB8EOR8/CmoI1BWsKa0OsDYFmQLNLRTOsDcHkA9+Ab1PFNwAbrDWgGdBsGmgGa22cTD6OnsXRs/ucG/uHK46ePSwNHD17uuzg6Fl/g8p4jp49ahaspPqNU9N37rV8YrU5zLjOyD/XLhJpdNtjKHqccwesrOCovThHbSsD9Z1V3i1MnZg6MXWeduq88zcM5voAIRpvsrJU3Ukkv8OQO+WuxlGO3WYBifJy0xYVidmaJ5Fz5NRnfbaJF9aBEA3QZKDJQPqfFz8G0h9oBjQD6T8pUEOIBvAN+DYOfAOwwVoDmgHNpoFmsNZGGaLRNZoG1NFoV5ksRGMwHE9BNzn0Tm65SUC3eWtthLany46dmd5RHdcul4EtTCCWFM3gWMtqE9rLGeuBXamtZrA3dNV37TvtJbkdlV1vNuMzPq0n1XNVtYj0RlHfWhdXt2tV59H5weBehXv1cPdq/XvV3OBfVwV7jNJY6PS79pK629XA0QnoSKtBa8W9v4jlo+yws27V5SwqdgHft6WrcnW1r7ZT8t8IHR6XZWnNQfPJJvGeP6yKShFhLn5usCt7uGjCOch64giMnbWjOMasQsgq1+c7HXTNAfoQdrKb8q1cRffRyusXZFexKy4I5bJ8eKDHKMiukLFUKWS749C12WG8ivSHiLiXxe5TVd+nA3D/K9zTyX/ag8/eUAA4A5wPB+fWF3yEbt0kmFY9J80W4VK1dH5oU+cOTvltXlTReojd/SZ5ac9ZY4Dn6oyUqcI9wtyBOxqUACgBEJznxwWA4ASaAc1AcE4K1OCOBr4B38aBbwA2WGtAM6DZNNAM1trZuKMtaWdZvLDyzzTJvhO7N+tMZAdb6BXmAzOZN5bI7GHRNd8ZOL6zl0o/YIKobWn1ZglH/g73btDxF0fHQzPYghY0g2GFwQrDmnIq5hfWlEAzoBnWlJMCNXgAgG/At3HgG4AN1hrQDGg2DTSDtTZKD8AzNIOhGew1O4QPW1tcvr3aH4QPZ+cgfAjN4DPRDG66BeG5hecWIsKYSzGXjmYudfoKEBGGiDB4M/Bml8qbgTCDFwBoBjSbBprBC4CYDeAb8G2q+AZgg7UGNAOaTQPNYK2NMmajazRBRBgiwqMVEa757uBshbMVksKQFD7F5ABJ4VNLCrfHOvAaeA2VYXAG4AzAGYABBQMKNAOaXQyagQGFvxr4BnybKr4B2GCtAc2AZtNAM1hrZ+OvtqTdRakMc2XW3/4pC6XKuIIrvXdY9dSVzHN2X6hUcrvtKQzCMtP7c6Oc06OOlJbFtiw63ESQjsVCAVMrplYsFMY4p2KhADQDmmGhMClQA60LfAO+jQPfAGyw1oBmQLNpoBmstVHSus+QjoV0rNfskLtrbWX49hpvkLubnYPc3YVLxzpPligKVWo3xL4SWSBTBxsdNjoYh3M1zsE4AM2AZmAcJgVq8A8B34Bv48A3ABusNaAZ0GwaaAZr7Rv5hwbnkVGNsna56auZtMdF8RkjqqSkJmQqY95xUkZHYHrFWmoJnWgje+Vc2ol9ePfhba0j1ZI7iop2xG6VBa8Srl267bpIvK9STbSFgeoveIJTP0SxzAkLNjKMhElHgYRpP06mEfX/cskedVElstkGT2uZOD2hYCkfmPo/hWqPV8S+z1ln31UQ3VeaPl+9X4U2fqTmR5R4Ozvymd7Dks9aez5qlemPeu1sbXZ52lkSm9zrHSrsDROrymuorl1Tm0Wr4HaXE5gdVyD5qF52aoOUDb8PUsqUN6uWSH9fP4mIhvw7KvNHsYtTEbbybq7rg3co+8K8FXBVBfzZnAvlOm4lrDg/yE6rZd2z1nj/ENfz8DTjpbHHJe6laLYn2e5jf6tUH1T2c1OJ7mH+4T3Mv21dLLguqMLtrFb7kCesV79cG1sGirggakRRFWdym2aFPgXoH6FW1Z0j3lelM2S3RR0lINZ/ihOnTL2fXVR2e9WZquINSKb6ZX7ByPb31NkW2DfQafPcLXW4QwZ4pcI2PK6b4+UsR6iXVmOQ1u5gnHaN01DGepA21i/JKpNk0FQGHXmfEzWf2t2UnJ5f5FPRjC9dTHCRXrKKoPW2V/7i91hVfM/1PrCe4JKYQZ1JsdHd1Bu+354DaVb2HvLj+IV5syXqK/PlrlClPN3yuadJksbK+XdvFBchVh8EtatfORJsGmY4uCSPDMlCNBqi0RCNxtForYn8rZphZaZXh6+sAeLFolX8H9O3EdmXW5HlvOo3D9HI4WtkFxVlvt/W4iUsL3t9g+deRDFfzsvVSspQhm2j402ZZQRp9C09pdPXameZ6ux8zaTn4VWtLB3V2DsZulo7FHy/yqEwDg70gsyW8bhwYL60G6dmFsCUgSkDUwamDEyZf7Qp47sRje9QJ35a3tFSQEP840fTHgdTkK7DHMpCOtrvLFlI7/EGC1m7Axayi4XkhENZqIF+XMQizk+5uPNT/Mx4hr+7Ss4n2knlzmYbxhNo+yF+DvFzX73OveD4uW9NGk0qcA7RwEAzoBmigScFati7BXwDvo0D3wBssNaAZkCzaaAZrDVo+0HbD9p+CEFACMJEQhCaGHfh2n5HzYIDZ3c1fGK1Ocy4zrrP6KLoD7vPt+90ruGVFRy1F+eobWWgEQFY3cLUiakTU+dpp847P+wt1woNHMlTWaouBut3GHKHRwRVUzkHRVSzYk6PjHbsNgtIlFct3JDZmidBgW+PaqEQNsI6EKIBmgw0GUj/8+LHQPoDzYBmIP0nBWoI0QC+Ad/GgW8ANlhrQDOg2TTQDNbaKEM0ukaT4x3t1+6zdOPYPN5VNklJQ7GkaAbHWlab0F7OWNuu0uCrb6p9as2IFtMEXPUm4IPY01V12FkPye2o7HqzGZ/xaT2pnquqRaQ3ivrWuri6Xas6j84PBvcq3KuHu1fr3/P2nHvXVcEeozQWOv2uvaTudjVwdAI60mrQWnHvL2L5KDvsrFt1OYuKXcD3bemqXF3tq+2U/DdCh8dlWVpz0HyySbznD6uiUkSYi58b7MoeLppwDrKeOAJjZ+0ojjGrELLK9flOB11zQJSs4pKMuCDfylV0H628fkF2FbviglAuy4cHeoyC7AoZS5VCtjsOXZsdxqtIf4iIe1nsPlX1fToA97/CPZ38pz347A0FgDPA+XBwbn3BR+jWTYJp1XPSbBEuVUt3iuR0NbXTt9Fv86KK1kPs7jfJG4P2N23QavU1ZapwjzB34I4GJQBKAATn+XEBIDiBZkAzEJyTAjW4o4FvwLdx4BuADdYa0AxoNg00g7V2Nu5oS9pZFi+s/DNNsu/E7s06E9nBFnqF+cBM5o0lMntYdM13Bo7v7KXSD5ggaltavVnCkb/DvRt0/MXR8dAMtqAFzWBYYbDCsKacivmFNSXQDGiGNeWkQA0eAOAb8G0c+AZgg7UGNAOaTQPNYK2N0gPwDM1gaAZ7zQ7hw9YWl2+v9gfhw9k5CB9CM/hMNIObbkF4buG5hYgw5lLMpaOZS52+AkSEISIM3gy82aXyZiDM4AUAmgHNpoFm8AIgZgP4BnybKr4B2GCtAc2AZtNAM1hro4zZ6BpNEBGGiPBoRYRrvjs4W+FshaQwJIVPMTlAUvjUksLtsQ68Bl5DZRicATgDcAZgQMGAAs2AZheDZmBA4a8GvgHfpopvADZYa0AzoNk00AzW2tn4qy1pd1Eqw88tB4bZpKsGqCxEFDvC02x2Cjmp48qvKjMpbGH9gjeqw3zRL+41ZfZO5dW6sqnANosfxS5ORdgqv7nOThWf2yZ3C+cl56pw9VA5u/VWLmou5tuHB65XtD1EOZdOp2iKatl2+1ul+qBywQ7w6vPmh/cw/xZPIiKwX0TJtuTe4XDCf9dDi/rlLo6dCxIsJWFnJrcp9T9XGS/cj0ZVTMIRvE9b1XJ+tMup1oNslvZ1E/r7lgp8SB/hDHL1HNw9XJOcZQ/xHm90ktqdY/vJCR1zR/cmVUv6zxcmX1M78VLnirTe7737Ow9lA4ZEs8k5Xmsgd8EvSSWVTbd7VLKFEXsxvvvnI/Wx9WtXwdM6Wq2DlUiMR1vq2Zm81al6kD4QRrQjFK4tLC+wvABZcl7rCpAlQDOgGciSSYEaXFvAN+DbOPANwAZrDWgGNJsGmsFaG6Vry5J7v3WQe96enDuiBu95+4X6o8yp81JiCalFB8td8D4WGxHcFmL1mUbeVuR5cHP9o3WTzf0USEV6SSNzkz5yQtwy1x/fOagpizKTc1CDgGfAM4zNs8NlGJtAM6AZjM1JgRqoQeAb8G0c+AZgg7UGNAOaTQPNYK2Nkxo8Ju+NQ538AZLLbPgAJxc8z6GHvUc0Dc58Q2GSLCZX6yT2QneYJN3uDZNsKLZd/WOO49gbjDkgowdGFbMaZjXY6Oc2ncFGB5oBzWCjTwrUwKgC34Bv48A3ABusNaAZ0GwaaAZrbTqMqjvaoYNb9Y9zGOZYM7lVA1SaEauHq2pXaq2uwx7aauXvo5yftvvaNexEiSqQIK0PZix5a/+zliE4yc53yHNfnjy3nxlvJndXVaE2Cneroz+GwW7IMyDyXHVvkdQRr36120fgnulxFLAqAuX2D3ndW3CMqkLzfTD5sKZgTWFteG5mFNaGQDOgGdaGkwI1MPnAN+DbOPANwAZrDWgGNJsGmsFaGyeTT6Rqul1kUuT0oX5aVSbhIr1fMCt+ZX9tHFeflgXfTj/LhKWs6/m/4fSDp/WuYgED+vJW9dQHmchM0KidB1qqQXXnTL4OXgW/3qpn9MtzLwe/vg7uXDL3URLl61o6WrpbAUFWRRLPu5MzRRhMcVkW9qe6IYKtUI0sqrDp4NWroMxL0pG4sm4IPuyW09VCtGJnI6JNfDRpyFKbRkkp2zq8JPfrnvU402ZJanVfK0imhj4dM0u9n+6qUoQPstCCztXx03ucG/uHa2+cuE2fT/DuWhm0+tcyU/1ykUuRrdbq8lMa32disxDxdk3autt1WqSq/hkLQrlg1Y5tJtW//jj6SxnFqjbv9FHrrVyRK0hkD+VG9Y3uFYvHPTeLfvDxu+00KLcP9eN22wfwdp7SWz/Id3jFdepv//I8qqpQ2eH/vGZc/L0TuU5w7nzVUxquyjfmFOxur6J3YHzwVqzW3lYI8qMENoVAV+eRpyo0NqjU8dvL5xsn/j/sUKXEhMI8A5zWR/NH5xD6k7d/o29Ly1Gz4MDREA2fWG0OM66z7iMgSKPbPNB/tDIctXDUNh21rQzUd1Z5tzB1YurE1HnaqfPO3zDIh4qbc3cqS9WdRvI7DLlT7moc5dhtFpAoLzdtUZGYrXkSOUdOfdZnm3hhHQjRAE0Gmgyk/3nxYyD9gWZAM5D+kwI1hGgA34Bv48A3ABusNaAZ0GwaaAZrbZQhGl2jaUAdjXaVyUI0BsPxFHSTQ+/klpsEdJu31kZoe7rs2JnpHdVx7XIZ2MIEYknRDI61rDahvZyxHtiV2moGe0NXfde+016S21HZ9WYzPuPTelI9V1WLSG8U9a11cXW7VnUenR8M7lW4Vw93r9a/V80N/nVVsMcojYVOv2svqbtdDRydgI60GrRW3PuLWD7KDjvrVl3OomIX8H1buipXV/tqOyX/jdDhcVmW1hw0n2wS7/nDqqgUEebi5wa7soeLJpyDrCeOwNhZO4pjzCqErHJ9vtNB1xygD2EnuynfylV0H628fkF2FbviglAuy4cHeoyC7AoZS5VCtjsOXZsdxqtIf4iIe1nsPlX1fToA97/CPZ38pz347A0FgDPA+XBwbn3BR+jWTYJp1XPSbBEuVUvnhzZ17uCU3+ZFFa2H2N1vkpf2nDUGeK7OSJkq3CPMHbijQQmAEgDBeX5cAAhOoBnQDATnpEAN7mjgG/BtHPgGYIO1BjQDmk0DzWCtnY072pJ2lsULK/9Mk+w7sXuzzkR2sIVeYT4wk3ljicweFl3znYHjO3up9AMmiNqWVm+WcOTvcO8GHX9xdDw0gy1oQTMYVhisMKwpp2J+YU0JNAOaYU05KVCDBwD4BnwbB74B2GCtAc2AZtNAM1hro/QAPEMzGJrBXrND+LC1xeXbq/1B+HB2DsKH0Aw+E83gplsQnlt4biEijLkUc+lo5lKnrwARYYgIgzcDb3apvBkIM3gBgGZAs2mgGbwAiNkAvgHfpopvADZYa0AzoNk00AzW2ihjNrpGE0SEISI8WhHhmu8OzlY4WyEpDEnhU0wOkBQ+taRwe6wDr4HXUBkGZwDOAJwBGFAwoEAzoNnFoBkYUPirgW/At6niG4AN1hrQDGg2DTSDtXY2/mpL2l2UyjBXZv3tn7JQqowruNJ7h1VPXck8Z/eFSiW3257CICwzvT83yjk96khpWWzLosNNBOlYLBQwtWJqxUJhjHMqFgpAM6AZFgqTAjXQusA34Ns48A3ABmsNaAY0mwaawVobJa37DOlYSMd6zQ65u9ZWhm+v8Qa5u9k5yN1duHSs82SJolCldkPsK5EFMnWw0WGjg3E4V+McjAPQDGgGxmFSoAb/EPAN+DYOfAOwwVoDmgHNpoFmsNa+kX9ocB4Z1Shrl5u+mkl7XBSfMaJKSmpCpjLmHSdldASmV6ylltCJNrJXzqWd2Id3H97WOlItuaOoaEfsVlnwKuHapduui8T7KtVEWxio/oInOPVDFMucsGAjw0iYdBRImPbjZBpR/y+X7FEXVSKbbfC0lonTEwqW8oGp/1Oo9nhF7PucdfZdBdF9penz1ftVaONHan5EibezI5/pPSz5rLXno1aZ/qjXztZml6edJbHJvd6hwt4wsaq8huraNbVZtApudzmB2XEFko/qZac2SNnw+yClTHmzaon0t/E3yI9iF6cibGXbXNdn7lDOV85BEfDXci6L66+VnqL2VlINiycREawsuGKHZwuvFHs827yxiBvAtNDC9gL7W6X6oLKTm7pwD/MP7+GuTP7iTU4dt2tzVc/92qAxGMNFU0OFQDaT2zQr9PE+3FbD0+ihcl97zkyiNqK+xg4h0+0ZF6oR8ILka45wL/XGkNrjXzqUbBjaT0Wu+1ozuQvdu6jods8GKmHiAMywP3brlH7tSuFXtFoHKx73pJMn9eRPwQSpetCA+Aw7q7B6weoFXMy5LVvAxQDNgGbgYiYFavCcAd+Ab+PANwAbrDWgGdBsGmgGa22sO6tWRoyqhQu+V4KowXs+1EH9UebUeRPtwIp3wXIXvI/FRgS3hVh9ppG3FXke3Fz/aMW35n4KtMFoSSNzkz5yQtwy1x/fOagpizKTc1CDgGfAM4zNs8NlGJtAM6AZjM1JgRqoQeAb8G0c+AZgg7UGNAOaTQPNYK2Nkxo8Ju8NvQ9/gOQyG9b2cJL8HHrYr97x0jBJPqK+1knshe4wSbrdGybZOAf+ql9o6IiAVafCwt+pBE34LOHBYEz3vO0OtVPewahiVsOsBhv97KYz2OhAM6AZbPRJgRoYVeAb8G0c+AZgg7UGNAOaTQPNYK1Nh1F1ms8d3CppNFt1g2GONZNbNUBl4qTn1S/VrtRahsuMksd0ZblWkq+uJ2Wlo+2+dg07UaIKxHLztQNDtYjE6xPYr4eqABA4vfvedj6nJmLPZj3yfXptfhoFEcqNuuGSDUyS8xPJh7Q/4BKtaqBvsFt9iSgx6jQ0YEheXB8HS19QKawpcpd1+GnW5lNoF+4UWr9L0sG1d3rYdUtCuDy2T631ZnEvM95M7q6qQrH898Kx74NghwNpYU3BmoI1hbUh1oZAM6DZpaIZ1oZg8oFvwLep4huADdYa0AxoNg00g7U2TiYfB9LiQNp9zo39wxUH0h6WBg6kPV12cCCtv0FlPAfSHjULVgL+xqnpO/daPrHaHGZcZ+SfaxeJNLrt4RQ9zrkDVlZw1F6co7aVgfrOKu8Wpk5MnZg6Tzt13vkbBnN9rBCNN1lZqu58kt9hyJ1yV+Mox26zgER5uWmLisRszZPIOXLqsz7bxAvrQIgGaDLQZCD9z4sfA+kPNAOagfSfFKghRAP4BnwbB74B2GCtAc2AZtNAM1hrowzR6BpNA+potKtMFqIxGI6noJsceie33CSg27y1NkLb02XHzkzvqI5rl8vAFiYQS4pmcKxltQnt5Yz1wK7UVjPYG7rqu/ad9pLcjsquN5vxGZ/Wk+q5qlpEeqOob62Lq9u1qvPo/GBwr8K9erh7tf69am7wr6uCPUZpLHT6XXtJ3e1q4OgEdKTVoLXi3l/E8lF22Fm36nIWFbuA79vSVbm62lfbKflvhA6Py7K05qD5ZJN4zx9WRaWIMBc/N9iVPVw04RxkPXEExs7aURxjViFklevznQ665gB9CDvZTflWrqL7aOX1C7Kr2BUXhHJZPjzQYxRkV8hYqhSy3XHo2uwwXkX6Q0Tcy2L3qarv0wG4/xXu6eQ/7cFnbygAnAHOh4Nz6ws+QrduEkyrnpNmi3CpWjo/tKlzB6f8Ni+qaD3E7n6TvLTnrDHAc3VGylThHmHuwB0NSgCUAAjO8+MCQHACzYBmIDgnBWpwRwPfgG/jwDcAG6w1oBnQbBpoBmvtbNzRlrSzLF5Y+WeaZN+J3Zt1JrKDLfQK84GZzBtLZPaw6JrvDBzf2UulHzBB1La0erOEI3+Hezfo+Iuj46EZbEELmsGwwmCFYU05FfMLa0qgGdAMa8pJgRo8AMA34Ns48A3ABmsNaAY0mwaawVobpQfgGZrB0Az2mh3Ch60tLt9e7Q/Ch7NzED6EZvCZaAY33YLw3MJzCxFhzKWYS0czlzp9BYgIQ0QYvBl4s0vlzUCYwQsANAOaTQPN4AVAzAbwDfg2VXwDsMFaA5oBzaaBZrDWRhmz0TWaICIMEeHRigjXfHdwtsLZCklhSAqfYnKApPCpJYXbYx14DbyGyjA4A3AG4AzAgIIBBZoBzS4GzcCAwl8NfAO+TRXfAGyw1oBmQLNpoBmstbPxV1vS7qJUhrky62//lIVSZVzBld47rHrqSuY5uy9UKrnd9hQGYZnp/blRzulRR0rLYlsWHW4iSMdioYCpFVMrFgpjnFOxUACaAc2wUJgUqIHWBb4B38aBbwA2WGtAM6DZNNAM1tooad1nSMdCOtZrdsjdtbYyfHuNN8jdzc5B7u7CpWOdJ0sUhSq1G2JfiSyQqYONDhsdjMO5GudgHIBmQDMwDpMCNfiHgG/At3HgG4AN1hrQDGg2DTSDtfaN/EOD88ioRlm73PTVTNrjoviMEVVSUhMylTHvOCmjIzC9Yi21hE60kb1yLu3EPrz78LbWkWrJHUVFO2K3yoJXCdcu3XZdJN5XqSbawkD1FzzBqR+iWOaEBRsZRsKko0DCtB8n04j6f7lkj7qoEtlsg6e1TJyeULCUD0z9n0K1xyti3+ess+8qiO4rTZ+v3q9CGz9S8yNKvJ0d+UzvYclnrT0ftcr0R712tja7PO0siU3u9Q4V9oaJVeU1VNeuqc2iVXC7ywnMjiuQfFQvO7VByobfByllyptVS6S/r59EREP+HZX5o9jFqQhbeTfX9cE7lH1h3gq4qgL+bM6Fch23Elac1yGtlY239Lb99FHF3ZqX9hWSU/+KWtyb6G2RSbF5sy6Tz63K0vfI2VzVxcp5dSpvsNmT9C9//nP3DhzKker59iAgEUchj3QR0wCS4ay2g0eZt3Gk1S6/s561XGHBplNXrj5VqF5clL5zxzr7mkX7P3d3HwP9NMu5zapmHULbtyS7F/DD5GfP12lWBHm52YhsZ8GXpfnUMI9Wa3K5UaGzSJCgqrZ29H3+DoNiIaL40C/rp+nTcZqosgXrciOSV6qpQrFUo9J7p5YdI2pVCFL/GvyWr5jKU5wuzEoQPFI0QViTz9qkGXnmNCDSlyuJQCegl65WZaZnyXq+junRpoVta7nK8/o4V1NP5WU1eVXKw/XHd/PgJi0f1vFONU8cp095cPPDm+Bf/8ef/5XnLPmboD7v9y/V1T2Xv8KTG93Fq/wc2PPNyPnnf+seOas4opG3Jutbz6mpejnZ2YQ1UgUPkRrngdikpTEnopr2LAYVBtU5DKp//rdalEPwgXq6GVm5P7T+X1rq9br8bSUlrSXZxldGmF66z4OPaqWhMlSo3iMehBoksbqZ2Znqv/XNVLnMHlXj0wq1JIuG5r4kKBP5m6pvCruSphowtDC0zmho/Td/vuKVOK1Pb3Vvr75vhtd1u8vbygjnwU+l6gnKLuRZaSnVxJOkfJxlWG1ON+RHi8jbkzAGFQbVeQ2qP3d97YAhxKOEupp4yKk0vCbrCFl+o1pTzWgiSOSTYQp0x+HlGLHTW5Gp9ZkazTn7NHWs64xv2/W/+r0VxbqTM6pkkM0CLw1W+psVN5Fm9KGq8ousVI3RGooVN+yy0SAiDsuIeekkWflFP6yMh7+k4a4R0/cSgHF8yp7gwkM9zUNiGhS3XeOT3QW/EisJDbrdo54hTBC44Xyej9TN0K9deRigRdKlZn4pkjxVDxoGbwZZDbiu4LqCI/7cfFZwxAPNgGZwxE8K1BA2CXwDvo0D3wBssNaAZkCzaaAZrLWxymqsjBJxCxf8kDSiBu/5RD/1BxP6kfEOxbtguQvex2IjgttCrD7TyNuKPA9urn+0ystzPwVSl1jSyNykj5yQpdYd1JRFmck5qEHAM+AZxubZ4TKMTaAZ0AzG5qRADdQg8A34Ng58A7DBWgOaAc2mgWaw1sZJDR6T94bYoz9AcpkNCzu6CFsOPeyXbnxpmCQpXNY7ib3QHSZJt3vDJFkus3a+WJ/KbPMDf02i/yqlTTNSReaNplZt12jvRrmnZpnTI4PBmO552x1cxZqCglHFrIZZDTb6eU1nsNGBZkAz2OiTAjUwqsA34Ns48A3ABmsNaAY0mwaawVqbDqPqDvzp4FbpgB4r0DbMsWZyqwaoTNy5Y+qXaldqLcNlRsljurJcK51dVE/Knhtk97XTsGc9wGrzPWn6zbXoImnCeY1rT0pqJupFmb67D+5or747vowG0+3tW622F+gkbeeqzhj7XvdPzs4PVJFaQTJdlTjaB1MXpq6zmrowZ8EQB5oBzaaBZjDEQZsC34BvU8U3ABusNaAZ0GwaaAZrDUf7fNOjfSxneZKDfbzEjimw7TE9x/p8b1LtqAX7QaoAK/u73HlCtC1C2eN+LbfsaNsuaVfHLhNb/ZCl5XYvu7uPtN3f37Q0bfs99Y2HTuK4u3o7spCUcexnwNLi3qWk3Cxl5l9p51uXtV2Q51+ej2ptLqgpltfK13z8i8raj1TJ/6mq/ZqeeO565KBmrRqsr12v1L9x7DoOhYCbQjTPdlI99D568JO0wdNGxtd1kYV+9ABI54fXaaTVrvs6ztVMqsahehNlkc68qryaJeqlpm/mjU6xWT//uZYsgMs+l1IHd9uUeC5UiZMy9srIZlC8ebCO9MhijV01ecZGYJfUNJKUw9W1IDVpr78rglBuZRLmQapNo3dqbspKffDSD/wcQeNKbMUyiiOqByfgS4nPu7rfL/6kZmqgt+W7CukKduXKe0UFprqbB28IKeI8NSLgO+uVoj5KJadERKUQTvfqzp+qyuf0paobqMbebIuFPYarv1lZ4flqdl8mXFOLQtDd7a5Yqx+x6stt9W3ff2XqOPcaaZ0+UdaMZeqi+/N58JddYGrzqiafwsNAFArEtkVVbJsAWawR6emrcSVCsS2qs890wwWvgl+pIj5ymX/g1+ZUrl9fB3f+BgPVHJk1mSkVts/+7+1P/xHoQdKdkl813SmqOTBO7RAO/pd94d/dm9Se/ztQ73d/wavt7g/oM8kovx/50SDfJYX4LXj1imTuDeDYr5naZpckny9mzzJbLKWy9yO9zOod5lsaRDzQt7FYyb3N/0bjEoGRafX0UWZZFGqkM40d6GwEuk9yFeiDz4wD9y8mW3P9bVUF11s9kikR9dZjxJazTsX6dk0XOPYbpljqIzf6r7w/GX1e255MzLmxGhmr6X/TbW2c/eHLF9cf1NeihJE+f37+A++1SULSXJd0AF4mjSaQaU/vaTpOgdZtSzpNIJd0Lt28Znib5hucEWuYraeMZjvry2WmDyzgCccA25H2lROF9+YKreB/XZ25pg+J4qm8rmNPdqHC1VT9nG1TPkiK1fJfz757/KfveIbMv/sijLL/83dGJ/+7L5XK/vN3hT6tTx9ik/M8x4vu2bootq+/+04ku1drlbgMX8UETK9yAqb5KtUkRS5Xyioudvzi97ayf/6F7hHQ3FRi+m+r0xAqMfyf7TYtq1Hv/OyN5epvRd32MW+ZnUxus1VrkVufsmwkgj4GwAsL+PlL15c9s9tPxbc4f3ZPV4cY2NeNgfhs3/FNEN/IsPNn1/xkp6FewPI7Np2jwambvuRPJbdmruROqxKk5tbd+fGfOhYR2tTgUeq/aA2DmsYXzUMGB/K/JUHjf2qalgzDKpe8DstZ91+lEmXKfst08AYREVpH7N5Aj0pUIfzfEjodI814amsnznNfLFVhCGB0rj7Q3KdhnobFRlnUzow/4ICMWk343WL4VdNVqLd+pyA04gNBDYelh+bPurb14FR/vBbVwRtmWNLV2jkY9iRLGoaUwpcvxJz8NYufn+myGmAZjT/1Jx8Ws6SmV/05jHL6O+wmYvyC/fHG4MqfGLUPP+Kjs7x2VZLQ+utRxCX9Un9+lruZd9AIDeITZHH48I+vyKRX/Rz6tJYiVOhEtaofMJTFqzvNjtgEWgeEUCH1G9erldwWe5/9xcPzjz/d3qmHl+YQEjLn1NVMPJH9of7lzKZcUXqvK137MotF8lAys6pxg0+N+f+CogIH +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Create a new turn for an agent. + + + + + + + + + tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."}},"additionalProperties":false,"required":["messages"],"title":"CreateAgentTurnRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/create-an-agent-with-the-given-configuration.api.mdx b/versioned_docs/version-v0.2.23/api/create-an-agent-with-the-given-configuration.api.mdx new file mode 100644 index 0000000..fd8ebde --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/create-an-agent-with-the-given-configuration.api.mdx @@ -0,0 +1,68 @@ +--- +id: create-an-agent-with-the-given-configuration +title: "Create an agent with the given configuration." +description: "Create an agent with the given configuration." +sidebar_label: "Create an agent with the given configuration." +hide_title: true +hide_table_of_contents: true +api: eJztG2tv27b2rxD+sg1wnKTYsK24G5C269atjyDJMFy0gUtLtMVFEjWSSuoF/u/3nENSoiw5sbPeYb3X+9DFEnne5/A8qNuRFqZSpRFm9Ph29OjoCP+XCpNoWVmpytHj0UnJThaitE+14Fac+fXsRtqM2Uwwji/Zi2eT0XiUqNLCLwTCqyqXCUcgh78bhHQ7MkkmCo5/2WUlALaa/S4SCxsrrSqhrXR0EMipTKOVxmpZLmBll7hfS/lHLZhMYYOcS6HZXGkiKyFyU0feaLUaj3iaStzG89MI25znRoxBDn/UUgtA+bZFfzkeWWlzxD8ggh4xjWy0sLUuAflNJkpHCRDPOCvFjSNoAhQhTV8OSfwiQxDAl7Hshhsmy2uey5QBZwXPgcEC6PxowjaW29pE6ySAXQjdY++ni4tT5lazRKVitGrEc5+aftAa1YKLxyAGkyltmamLguslU3NSmKA1N5lMMiYd01pyMC3UKC/9e8KzQviWy3xbzG41os5VCbyxrC54eQCKSfksFyza0yEHMckSeC6TLbj8/E3l7OsLdsJ+PXvpmUmA+JlgtQGDsAqNQ0txLVihNFhuifokzTE+U7Ul5KYSCZhzwlSS1FoLQL9G1y727DUctNUILzJvEtMG4YUQweZaFUTDyemLCTtT9SLLl6CePFc3hp09f8q+/ubo6wkKTXzgReUsI9gXmHqL7wlP2Zkz8ZaeLS3fe86jb4c9J8klBqQMthv8wyoFm8tlAIxQwRAW8hp8kxeqhjUoW1mIvVPtneoTc6pH37b4LsDSX6Gle88ysWv9W9XgEyAh8SERIkWxobPBecZyWUg7Yae54ECQBevhCw5OksNLHU6qrzadVEboa1A+iBNdSWg8c0tWl+IDyBtPYOHFsHetvWt9Qq71VXxevUDLBsrZubP2Fr93r5O+yQdhpBP2pgZLELygU2km4OApFeWraXCvVMx5ndvB9Hsj4L1T7Z3q03KqoyFsW7gQeQmaGl8Y5IbKMYPEd4l1FRpaiitLmyrVpXvgLXO5qDWJGgvWimteCPBthHrpJAYn5xOVLpHsj1rPOuQbktaYsKaG9aXi+D7nRTGDLU6JG7ONt+O5v1huONA9OBaWIQWqFG/mIKP7QLu3fScBBiHVfzxaaEg+lqMo4MWPYlJe1VAEgMu88wvejdBxfKW/BPlAfOiRuptPEJGRA/xIiM490PMAc50yt6yPHGjiENFEDoIxpMBMLjKsZUBGMz6TubSwSF1h4WGZ4BAXjBUVOstflatV1bTqiLV9MixVer+lUEFEogBqwIl1TEZZF7OBQ+IpuI1WuYH0skxVUQpDRVcAPGE/oVw0g+IOvA1eYYsE+ybNckJJDAwi8zweTb79qoe7LmpIXMHdu1LPIIJlKk/Juco6yUVtIoqeOZgGxYFg/5ohXajq9F4zgkUHFfvc0/LFRnuiIGrawGugFM7RqoxwlSsalHGhLhnm/vvv2EezsquelV3dY2VX21sZQbs/C3lNpuCYr4IAAAXSCWhcF7BVbiAHpIk1jmXHD1BvoK6r5l+2UvPVJuWCUYKMMVw0C4AN1PIVpAZAaS6vRO6DhgEV4oknEXwhS27hAAVheR0uX8NRhipx1BZwYqH2YIGPsPeeC/vg/bDg3QtV+yC+D+L/tSDei5L7YP6JBvNVWzyPeqhXGMI/TN3iDWIMLtQjE3DCZlnUBSsj8TrRIq2+HoUCAwMC1Fvg2jQ2U2ioVB4xhOICXhK65EtVazTKooK4mIPztSQiyFJZ318kYIVKRf4ZNgqA5g+or3JhM7JhLSrAgnimlQCl2eWdQeJ4g9HMhL3BdsrBo8kRVH0pg/9P2KkyktzHxybCIP8UNH/z1M44luJQbN1kwmIcg3+WDM5MwXUQBhFtFJtzPQ4BjrQa8UY6lZlSVNdbnl/5Et7hqmRiiGEDf0cscq05mpeEmGz6XrrqDTrRM9iXEC6gQC0TYAro1iJU4hA48pwhjqBSpHNea8eaMzjmxit+NknM0TbUGqqIe7YJTEA02bXN4M33sm/ap648XWctvGZtMT5xPZeqtlOTSZGnZgfRwVaQ/4P3WqXyhVZ1dee+Xj3cKO6+sFxSljagcK4Xg7X7sOwHSCjrPI8JmAEngpfxI+9Z0ZM+3Y7XPiOry9VOpkCMerbWJ+kXQNqPKOTf4MA7wRWroSWkEDfQm6JetlHlTmLvhRVYFZpslghYW7FD+++nO9uLNxiFAxqWKkHpUtyP+vs4bbB2SJhuSiDWzhp42wPEPheTxWTM3B4MnnRofbGzPLcUYos4SLK1xJ4/ROeK1bVYR/lbcyBAZtMChh8BpmvQod5kea1cU3DUbeL//c55hyX6vNgdh013MWJtTmcAmNE1JFzpbhHfu/ma2XSpiXZ0Mi6Vnzamt07/aatSMZelbDqjnEQ/6Z+REccvpaF8pfUnp05SGk8SUVmyEngHlmUHe7j/oLh715ChoZIFZqIZgg1BdDd1rikJ7GcgE8aw1WqGJhuQQGjIFGUhQs5nJs2hOk0yJYdHKALkQzevaqtiYxmPSpD2OjlPHZyNfouUATkdj0W7AeA4ZEkwEcLhkgFTyiRdEWlTOj+vKfiSfGIm/GwDx3gvLHAMyWRqMHHELS+glNJ1QhJ4TuswkUl45Yo2EHIIjQR8QkRXWiSYc7vwE6TjsuqpmwTdJSSaPoxH87okvFPL8W21tBn8yMHwez5GkJ87wD03c8m8Q0v+lUBlSlxAoQXV6iH7U2iF4zq6S5Obu5hoZxz31KZde9iUTf0/WcZAbLiMi3MvgY2BYIjJhrFxw+8YGUbZTdhTHJPlUN+4SePSR1aGIQA5RyC8HUNS1O30WFqRTxCTb058dFveyHKQsYmUlCksu4I9Ez9o0FD9PFkyL80xe5nzgrNzy5MrVwNxi80x27IdAGCGIXFoDwcKRNfKugFuqzh2wN6ve9kE+Xr/2FfQgQTGtfBw6WYqWtnP529eM+ckw5Bi0QxDhCotVz76cvavsOG7Zifq83soTRfDGCJpDyNw1RTSe0pLmVlCsfiBHRzgLN2ftAGblzaVvEsDQp0Wwhi+ENOZyPi1dC3rjW5eoRORo1c5T8Sd6n9KwYZilte6uhZaQwpD6vHKZo4M37UgEZzTk1eOrieerInDDSI4qZwnU5Lkc6IAxTMTTGBXHJ4tQHLm/jKbwbTD6w1EuIJ+jTBs7EiwiNpLwYmYfXZ729hDe2qb1eoz6uaVKc62RdRUaPQZrcZDoM5TDH2yNBDN8f5KJ0Y59d2bcHRitjsyBtrP6xNxH9gmoTUmyzmkm3KtWhrojx0f4Q701i3KmSaSRJfKfVTozOXDtRAf5Ae6CUOgvbbijX2w21VwwRFcvO4AGTfpmBW5gLwQ7/SVacNSW7KIEuuqqQHrcb04bcChRfeqy0Dh5BW5gaJ5zhfBqND6Qo/No2GUpeJdKBAsGFPAmrrCzd00iU6PDVjaOykuVncuUXyEOwvxITX1dz3GowVUFAXX/cjU9smjDUPd8nC0uDoptOO7WFqfip+DeOKfn1bRguZPx52j3t0/DSr0kQVVCNIBmUFOhceZP3HOn/0ydsKTmDdZHMKx99Uy5SDX5H2TQz1gEBALNEonf4bH5/Q0fNaxIYHuB6qIyYNFTYG74TN0Z/H60cPn8Q+2TL9ke6sMMLszX/cMhD0r55+gFT55/Zx5Jpr01qnvbpt8kH2hiDpjasK7s1F5eu82qAfdDrgnouxj5T5W/hNiZWNMf7+N7qPm/3TUjMfwaxh2It1VGGtVQe8bUlfuPOCDVN9biwC6284ENnxMR8y0e31vDiqATMHPUaXog7uK2wx+HV4fH3J3kxqqdfqywZCd1RovZGfWVo8PD3m5PMgUpucHObZNDgy2TSaJKujAMSKptcT5PWx8Fsz97SW+Q087a69U/9BeCF+/Ej1whzm+pewN1N82W7sacTR8neA4DN3fBtu/7E+V23f9qXH8Lp4KN8/HYdjWxhg/xF0h993B5dv+4q71No87l9L7u9YHdPfCWzOHy87Ygyhd793G3dnQeBxq7IX+3cZOT9MRGKjYj4+aoryltVtRt8/XhXBX3dp4UK+Q9ALrHgDdLMF9Z4BfZNAG72lxv/A8jn2wH73GSfz6eKDmd+dw/EEHD+NLfBhDpmajb/aYdyVb+4+Hy2Ig0UpJcFsq6gGK1BBvtWM4xYIbmyB67vtLAJRr8Q4veORKU/+yD5wanLkAZrCL5Kh6hdpxvTwMHAUv2/nyzp9arI19m08rdgbklYjTpcMq55L6GBSxbn1ge+s0wZuPRDB84ePbW7z286vOVyt8DIFJL92HH/Sd0Qx1/RajRCZ4Cscnut+VWLqjBck9uHDHE01SnXl3vwrB09TtOKEp451rL6PIfPrm/ALPPf/lCToGPNX8Bh0Y/n08wo5G1bSa6NntKOflogZGgy/SdzL/AdklT00= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Create an agent with the given configuration. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"input_shields":{"type":"array","items":{"type":"string"}},"output_shields":{"type":"array","items":{"type":"string"}},"toolgroups":{"type":"array","items":{"oneOf":[{"type":"string"},{"type":"object","properties":{"name":{"type":"string"},"args":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}}},"additionalProperties":false,"required":["name","args"],"title":"AgentToolGroupWithArgs"}],"title":"AgentTool"}},"client_tools":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the tool"},"description":{"type":"string","description":"(Optional) Human-readable description of what the tool does"},"parameters":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the parameter"},"parameter_type":{"type":"string","description":"Type of the parameter (e.g., string, integer)"},"description":{"type":"string","description":"Human-readable description of what the parameter does"},"required":{"type":"boolean","default":true,"description":"Whether this parameter is required for tool invocation"},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}],"description":"(Optional) Default value for the parameter if not provided"}},"additionalProperties":false,"required":["name","parameter_type","description","required"],"title":"ToolParameter","description":"Parameter definition for a tool."},"description":"(Optional) List of parameters this tool accepts"},"metadata":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"(Optional) Additional metadata about the tool"}},"additionalProperties":false,"required":["name"],"title":"ToolDef","description":"Tool definition used in runtime contexts."}},"tool_choice":{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model.","deprecated":true},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"title":"ToolPromptFormat","description":"Prompt format for calling custom / zero shot tools.","deprecated":true},"tool_config":{"type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."},"max_infer_iters":{"type":"integer","default":10},"model":{"type":"string","description":"The model identifier to use for the agent"},"instructions":{"type":"string","description":"The system instructions for the agent"},"name":{"type":"string","description":"Optional name for the agent, used in telemetry and identification"},"enable_session_persistence":{"type":"boolean","default":false,"description":"Optional flag indicating whether session data has to be persisted"},"response_format":{"description":"Optional response format configuration","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"}},"additionalProperties":false,"required":["model","instructions"],"title":"AgentConfig"}},"additionalProperties":false,"required":["agent_config"],"title":"CreateAgentRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/creates-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/creates-a-vector-store.api.mdx new file mode 100644 index 0000000..87c7169 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/creates-a-vector-store.api.mdx @@ -0,0 +1,68 @@ +--- +id: creates-a-vector-store +title: "Creates a vector store." +description: "Creates a vector store." +sidebar_label: "Creates a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztWVtPGzkU/ivWPLUSBOi225a39KZFahfEZaUVoODMOBkXz3hqe0IjxH/f79gziZMMJBV9WCQiQTIzPld/5+Izt4kRttKlFTbZv01e7e7SVyZsamTlpC6T/aTP/hGp0+YEf+Jw+B0XzIgKhKJ0shwzlwuWGsGdyNjEL2WW1vaSrSTVpcMy4sqrSsmUE9ed75ZY3yY2zUXB6ZebVgLCtOcPwsroShgng2Iyi9ZYZyAWaxb1PCvlj1owmZFaIykMG0ETUi5WKrnbaoV0chzxWuFREmgGgWZZVOMFoo7kbTGubvjUsosF6ouEZDYOGvBYroRzxsKs8D+VhbCOFxW7yUW5YgO74bb1OPEueSHW++fFof/F1Uv2NwiYHnU6p7Z8LAbDqQue79C0cdHusgACCGiZrXgqWG0Bh+F0VXlZssAdwkZSiUGq69LZVeB9wUMGHKTCWsIZPOJqGO6Xd2/u1hocpbqolCC3rd+Ev+tiCBDBTaSlhTDuWM4ngg0FNsXWKSk2qpWatlqG7Uh5mQqlHiMkI9Okia2fcyW/cfkI/oGaOd3yJ5ayHOBqbOjy1/mmtTGIA7hiKEjbBYc47bjaBPa0jpVLrOVqACR34MqzTAZAH0V7POLKii2ktR+1NOSi82jT472ZOXHR9FbbS/yQTpG6UfojSH4McIUGAZAP55FY+KK1H4PLWljfE47iZwU77ICPHDzWkSi73XCb6FIcjmD+jKIEUsFxdj3UWglexreC6+M7jUXRHW4Mn8Y3Gk3uLu8eyDifyQyf+1mlUQam9ybnmcGbJMpIwtqcKZVigTmJUdy6AU+dnIhHiMKmESPmGUk3ZboLrFtJIRzPuOusdP/jDTwRjky8FtPtCVcorhWXpkkiiCTEOuPOcRRxn00eF6aSQmTmlaheLtakxaIxC8LIx92he9hyXirllSj7B02Lw/xKFpToQX0y4HVXU3QKU0l54MBXY1nCQTJj4FFwBWwXId38nvZnJdHcB9K/Tk+P5nUy89hrfLGuPfhsDEUkLUYjw2yuDXJTXRTcTNvkJPyam1ymOZPBaCM5chgFM/AQnns5HksOGXZTyWE1iVa6hG0srwtebgMGGR+iEYhoFtQJtQs2I7H/UhPUZ2fHXxtjGjD7ngVANgLUApW+CA0L7WdIXnyoa+eF20qkaPtSptNQ/FKxpNevgH+G4rBbM+dFWPZuusd5bRfPRkYXXof+0UGPHet6nCvKtUrpG8uOv3xkb9/tvu2FRMupNMX4AtTn8j6gCzkOEJ/rsyHym8h59b47clIlqfLlIKdDBFyuQVxOW8a+6HM2RnLGd0Gh7n2LvPscVM9B9cSC6tX7ubxTIP0bIb2JLBuH1r+6DgcM8TMVIhP+FMDQNgmmZCFdjx2h4EMhB/TwMUeQKDw0baV6c1+lssJMsPlwJ4WSgHsIWHWJdgiRgSvRuOE5tJ5D6wmF1pu4Xh0QsqE5Owlon8tvwqu/CvnWGVmPHdZAguCFr0r+hF9qP9fJ2vCaHepWJmQPMH4OquegelpBtdslbYMQ8lFCUONjS9aEI9XBIam/NPnwZzuLXV4eF1fccJzjhCEOl8E7qJIfdDYN47vfEkmbzUv7jNZ1zil6s9mlzOKIDEdrQNGJomMutHK+7qOuW9/a+lnnwafmbL0yvEAg1iojaPbYmRWjWgW9tFYWPK4Fu/LqWMFNml/NT+jcDynDJK331OdJ1MqIzkHSMpJoCpvX5TVkDSASaBtPn6CxrQ2stWGWnPwTjxPa2hf2ZY8djKhioddzW2HYVVOniBVXvHb6asaj9/QHUnt/bjyTwp3ZOAfohzawDqgoUBg3KEseci0R80TElVwbEgMq4Ar05lIynFhLK0N6Wjf9hqjZ+lkVmckOUix70bQh++yPd69fennIbhOZCTPY5DUZiTn41PJvSdcZta6MtRWDxmlchgQfDd/aOYIvEPOS50wtAhhzjcuk0n7WUHGX42pnsrejPT/6Fb9S83M/3+VZj8faUHHKnav2d3ZwutrOwUhk20rxgm+jsKXXvVQXQBTRoVxJN/WEn9qW7vySnlEhOZ6XnM/z4hhKxtyn8+R/3t68XE2u96ShxfjrguVcUCeWdpf2PCow1Kd4EDQb8pVcwE7IBeyk6VZ8+gRv8l8AxWSvAyfU3NmFNofHjVfM2Udfo5K9KNnSh1NWIjpRZpWW9PqOl+07rlSb0Ll4mFN0mBFP23dV3IiLkroPbXxArzIf0iRKoVHzLx+DVt/IjaHgEabQQ0abeH/7seCBqN14gKQJNid+up1K4Uju36Ean1oCjs+DdwOSyet73vUxmgEdQiytvb0d4oR/ZtTdHd0GFs009EK+zR7Spp5TYszRExPIgGLkQtIxaLt9ShrRciRHyq7LjRIl2kDRR3NQuQfXXkbReXR4corFw6YZI6jiruE3uEn/9xNvJlH7euHv3SaKl+MaW4PngSd9/gPTymit +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Creates a vector store. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/dataset-io.tag.mdx b/versioned_docs/version-v0.2.23/api/dataset-io.tag.mdx new file mode 100644 index 0000000..dae7d43 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/dataset-io.tag.mdx @@ -0,0 +1,19 @@ +--- +id: dataset-io +title: "DatasetIO" +description: "DatasetIO" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/datasets.tag.mdx b/versioned_docs/version-v0.2.23/api/datasets.tag.mdx new file mode 100644 index 0000000..2cf7b50 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/datasets.tag.mdx @@ -0,0 +1,19 @@ +--- +id: datasets +title: "Datasets" +description: "Datasets" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/delete-a-file.api.mdx b/versioned_docs/version-v0.2.23/api/delete-a-file.api.mdx new file mode 100644 index 0000000..47f81b8 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-a-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-a-file +title: "Delete a file." +description: "Delete a file." +sidebar_label: "Delete a file." +hide_title: true +hide_table_of_contents: true +api: eJztWF1vGzcQ/CvEPSWALDlGgzR+U2sHNeDAhuygKGyjoO5WOiY88kryZAuC/ntneTx9WDLqAH0xIL/oPpa7w9kZ8uhF5sjX1njy2ekiOzk+5p+CfO5UHZQ12Wk2NOKqJjO8+KI0nZGmQKM0SChTqFwGZabCN3lO3k8aLQoOwuB+1styawKZwGllXesYbc3gu+fci8znJVWSr8K8JlSz4++UBwysna3JBdUiU8VGjA8OFRGzDfS2JDEBRqEKVFQTRU6EUgbxKH2LiYps2etK7MkHsB4vMs4S009ko7cePK/X5hKcqCceS5WXQnkh9aOce3Efx91nXLSrv646tlaTNDtp/ywplBF6mg7DX7Or5+u5IK8sCsUDpb7eIGwitacemvtPoxxXvWMGe2t2uxQPvSyooBnPS03eAbjq/sS61Gv0XybuO7UIzuTF8PqiD6AM9Zd96mISGSb5tk/KzKRWhUDqSmpUqKj4/2TkgwyN34hTSDsltzPHP25vr0UbLXJbELcwEfVfMjx3DuBjcA+s+NK6gP5VlXRzYSexrxRjVnrhSTslTYiUSpPexzpRO0Eq/drKbTSX1tZgbqJsKmmOHMlCjtGhjTFbcLiSgv6lyV8xy3dXdSu792Iovo0u02RygB+TaDwVIlh0FqNpRqKyjrXB/YydE3JsmxCL+5pymDUXNs8b5wjln+H6GZmnDnfdWpG3IfRI0wvkuZW2na0iBhawGNlmWsJ5E6u1ffRi9OV38enX4099Jo2eZFW3yuj0Bamv6/0mCzFqJb7G80rlJ+ecfN7vnFwruEKUvELwRbAWg828S8xZIYSpmhF+K9sghrlVFR1MdTDVGzPVyed1vVso/SsrPTnLb1rrL9vAE2CInnKigmljs8lAQqtKhb64xtYLQAHqkVMJk2i8dN1O9fGlncqTm6H5oJOtRKCHhdUYegLf2E8TuQdrHaz1pqz1cXO/umBlA7m4adW+rp/sNdyVfEdG0RdXDZRAsoq70piw8Rgbv8aLzl6r7+o9R40XEx9MdTDV2zLV8b5qr7BQdAlLTU49zyYepBj7Ntb2lJbOXXzUrqWTFR45HrXIDG7S0fXvePhTPKqWodx7kr046wiP5zj0rUlHvFBCnemDkuusmQ6uAfM7vktqWS6BGYBKi9B04owwgeA0G8w+DNBEIxVfcU0/WCSwS4S1W207lcYxZ2UI9elggE3/qLQevB1pLSt5BL7zH/3cVhkX9AQWVZjHgWfdSnP3wO9YexFnassljxc3PF7cJAVGZaI+F2/ZmX3YQxgb1m9JV26aaTNz/B8ElqCZQpJ7I579YXWg+FVOpqgt1iIP88cvFuWw7LhWjQWfsXmhchOZk2+TSkf3hhVlXbTabvIxny40zOfktEP1FUuZ9lG1NXjEusCcJLnsqGpr4ov1Grwbmbof6CkMao2PKq4QW7dITb9ruWzbzhzzXWw9fk87paJR3F8OXyzG+Ez75vRyyY8hQceNxWVcK8fcRbS5UJ6vi9Vi8CLod6Mk3vfi50S/d3LpIRTJk5G64Ttc/qD5hvOWDwgtsfJii2G07dthnlMdNsbtbGlb9jk7vzy/PYen/gVqgdkg +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete a file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/delete-a-prompt.api.mdx b/versioned_docs/version-v0.2.23/api/delete-a-prompt.api.mdx new file mode 100644 index 0000000..1512894 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-a-prompt.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-a-prompt +title: "Delete a prompt." +description: "Delete a prompt." +sidebar_label: "Delete a prompt." +hide_title: true +hide_table_of_contents: true +api: eJztV8Fu2zgQ/RVCpxZI7DRo0a1v6SaLLTZFgsQ9LNJgMZbGFltK1JKUE8Pwv/cNKcd27BS+BnAupsQZzsyb94bKPHPsG1t79tlgnp2enMhPwT53ugna1tkgu/onWxxl73dtDUtWjv9v2Qf1QF7pekpGF8o6VZEZW1dxkR1lua0D10H8qWmMzkn8+z+8HDLPfF5yRbIKs4ZxrB394DzAsXG2YRd0ys4HCq1fs9M4dsIOhptp/T0cXqtkrXJbsOQfdDC85uuD0/Vky/XCOSQfjY8UKV9aF5Rvq4rcTNmxCqiYo81DqfNS6VS001QHhYIV1d1+jLOQ8wNps2/kZC2hja1Rmyrbiupjx1TQyLBa89lIRyLpGjXX+R5VvrmKKzJv1Zn6dnPZFZMj+RGr1nOhgkVn4c1TVpV1jDKln7Fzika2DTG4bzjXY50rm+etc4zwz/JCYlQUOsW7XmvomIzno0z4ox14MrhbdnjZrSfw7p/6l2B6Abwll9XY2SrmcHb9padubDspzQztMcY+eHXz15/q4x8nH3sCGj9S1SRmLPkFqq/ifaZC3SSKr/LZk/mLhVT//vTTbuXkRkMVqoS7l0WwFs71bHmwnAoiTPSU8VvZFjaCra74IKqDqF6ZqE4/reINwfSvwvROWX5dWv/aFpoAQvyYMxcCm4iNAiujKx166towIaEA9tCEIBKDTdfr9PbhpZvKs5ui+YBTpMSAR4jV1vwIvAOeuIPhIK2DtF6RtD6s31dfhNnIXN0mtq/id/I626b8Eoyip65aMIGpirfSiHHx1DYAMOx18ip4TK0J2xL73cEHUR1E9bpEdbIr2h4SiioRqtHESzUotmpwwd0/z/acDeNKI9VEi54wnRxVeOnEc57VeIBh2v9Pi4i0eDYUyq3i5YLTBQQmYnVL+JOv9LGI4STKCuvgWmC/pbyOL4sFckY6pYVplvxjkog+yPrTd/10uu/Pn1JcwCBds6mE1gleZQjNoN/HhX9cWg/Mjo2hio6Bdf6zl9sqk1CegaAOs+h4vpwyd/eyJ7yLGXYtuRR/dSv+6rZjX2Ql4kvwhMn03Q6YRKx+g7a0LqT1k0NJQSCcAlj/vVbP/jAZOH6Rc100FnPIQ/jxa0U7jByXmFgASyVDyo0pZ58OJcffa2GTdVFm24eP5D8LA+E5miyz+ooxZnxkbAMcMRMEk44mO/i0Ufp8NYF32Xa9D/wY+o3BR5VEie2bdy2/S3h2TcdqsGImWiSdFaP5fISPs2/OLBbyGt93TlqKZZyQI+kfGlxoL+viaQS8mOybm46wb9X+JN9ZTvcSPBSakGnlCcufPNvQ2eIexiWmLa4VyTXtn+U5N2HNc+sa2xDM+cXlxfACKvoFc4kO+Q== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete a prompt. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/delete-a-vector-store-file.api.mdx b/versioned_docs/version-v0.2.23/api/delete-a-vector-store-file.api.mdx new file mode 100644 index 0000000..76d34bd --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-a-vector-store-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-a-vector-store-file +title: "Delete a vector store file." +description: "Delete a vector store file." +sidebar_label: "Delete a vector store file." +hide_title: true +hide_table_of_contents: true +api: eJztWGtPGzkU/SvWfGolSABt1S3f2IVqkahAPHa1oqhyZm4ybj321PYEoij/ved6ZpLJA0pVviAFCWUmtu/j3HOu7UwTR760xpNPDqfJwd4ef2TkU6fKoKxJDpMj8S+lwbor/NNHpemYNAW6bBYKZTKVyqDMSIScRMajWCl8kKHyvWQnSa0JZAKblmWp42xr+l89258mPs2pkPwUJiXBox18hUcsLJ0tyQVVR6eyzhwfHDxiznKwN0Z9rxBTBn9qqMgJO1yERZkYIoFkttP62GhwKCuNoWQc8/7iOfEeL+w1Vtbcnkdrgk11fQ+tW8akRZsjaG0tQhhYq0mabgzBVbTq7L+cYHTFMgMVYRX30gtfpSl5P6x0MoMrmWWKx6S+6CA6lNrDuKPvlXIcyC1DvLOAv43wbicJKmgO8UkqrMEy58jQ2aKJFTSRokZWRGRjSXqIkyP9YxMDr5EoR0k+xOyUGUutMgELhdQAuYgleSGa1bztzFMwOyK3lt0/19cXDctFarNY1Aann9H0xDmmBk/eARw+ty6gaEUh3aRlLMU597lKc6HqpJ2SJkRWSdOMRz+RTUEq/VzP9Wx2ra1BbiKvCml2HclMDjTTar5mKRz2pAxyNukzsnxzXtaseyuOxM3lWZNMiuAHJCoPOQaLymI1jUkUTAZluJ41k+XAViE69yWlkFQqbJpWzhHcr8T1KyxvKtxWaw5eh+cRpkfAc0us5hiOLk574tJWo1xPUB6t7b0Xlx//Fu//3HvfY9DoQRZlzYyWX6D6wt9fMhOXNcUX8TyT+Y1yDj5sVk6qFVQhcm4L/BCsxWIzaQ2zVRBhpMaEz8JWmMPYqoK2otqK6pWJ6uDDwt81mP6Jmd4oy3el9b+toAkgRA8pUcawsdhkIKFVoUJPXGAzRkAB7JEjCZFoDLp2p3r32E7lyY1RfMDJUiLAw8SqDD0Abz6CUAPDVlpbab0iab3r7lenzGxELq5qti/8N/I6Wqd8C0bWE+cVmECyiLvSgLDxGBvPzFkrr/kJeO068oThrai2onpdotrb5O0ZEooqYarJkeds6nvZ6TmHvxxufUXbfOcC66WTBcYdW5kmBi8rF98v8U6o2FQpQ74GB295p8dtIZacsBaxb7a/DLBPrnV9rWTvi3rU99w1dTacQqbz2NjKr8X0O34BJ+DJLaY21+EIGnweJv3xfh8UM1LxUxcz35+uQDjrcxT4vgl/BjP1QaEGvnJc8TyE8rDfx5FlN7ceVd/VWhZyF2xJv/VSWyQckCdwQIVJXHjc9snbOx5j5cQ8GlKd8XpxxevFVaOfqCv4Z+c1XuP9DRByu/FLwpPdVtC1HHIZBBroWMHIZyNW/tDbKN4pyGSlRSf1aF3xvKUcSOJqLWXME26zbihT8rVR6egzk0hbFxvFuvEB3400WoeTozaqT2jE2kfNlcARXY0xaQj0tCCWUJgutpOfLGt4E+gh9EsN0rPvWNRpQ5fbGuWaMIz+fixBhzR4P1xXXuQND7XER5WZHGxxOh3ghHrj9GzGX+OQ65gVeIzbxIApAI5kyvNzNu+Djyb55rJRxlvxO7reiEbzJcjNiUtd8Rsev9FkQ8thxb9A3C8SW4v87A5Tc2yPOAcwrvXoUZpSGTrr1s4dS13k+OTs5PoEreUHtyqAjw== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete a vector store file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/delete-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/delete-a-vector-store.api.mdx new file mode 100644 index 0000000..265fc37 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-a-vector-store.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-a-vector-store +title: "Delete a vector store." +description: "Delete a vector store." +sidebar_label: "Delete a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztV11v2zYU/SuEnlrAsZNgRde8ZUuGBUiRwEk2DGlQ0NK1xZYiVZJyYhj+7zuXkmL5a/OAvQRwgMCyxft17jmX5Dxx5EtrPPnkbJ6cHh/zR0Y+daoMyprkLDkXf1AarLvDP12QpkDDxkgok6lUBmUmIuQkMn4LK+GDDJXvJ70ktSaQCexWlqWOq60ZfPPse574NKdC8lOYlYRodvQN0WBYOluSC6rOTGWdNT44RMSa1UQfjPpRIacM8dRYkRN2vEyLMjGNdSA5FJIsem2srY7HstJ4ldQ2X6NNv3G0EfkmOhLspRt+jGArsLRgc/DW1zL6yFpN0nTDB1fRerA/c4LTNc+MVURWPEsvfJWm5P240skCoWSWKX4n9W0H1LHUHs4d/aiU40QeGeXesgNthk+9JKigOcWdTNiA5JUiY2eLJk+wRK40oY/0OMGftvHuHvVxcuRDLEqZqdQqEzAupAa2RezE/0SwmrGddQpuJ+Q2Cvv9/v624bdIbRZ72cDzbwS9dI4ZwYt7QMLn1gX0qiikm7VcpbjmOVdpLlRdtFPShEgmaZr3MU4kUZBK7xu5Xs2htTWoTeRVIc2RI5nJkWY2vdqspMORlEHNJt2jync3ZU229+JcPAyvm2JSJD8iUXkIMVh0FtY0JVGAByiT+1kTWI5sFWJwX1IKJaXCpmnlHCH8Wl7/hdxNh9tuvYLXoXeEaQd4boXQnMP57VVfDG01yfUM7dHaPnsx/O1X8fHn4499Bo1eZFHWzGj5Baov4/0iMzGsKb7MZ0/mN8o5/bRdOalWUIXIeRrwQ7AWxmbWOmavIMJETQmfha2whrFVBR1EdRDVGxPV6adlvHsw/TMzvVGW70rrL1tBE0CIXlKijGFjsclAQqtChb64xR6MhALYIycSItF46dqd6sOuncqTm6L5gJOlRICHiVUZegHefPigBoaDtA7SekPS+tDdr66Y2chc3NVsX8Zv5HW+SfkWjKwvbiowgWQRd6URYeMxNh6Vs1ZerwffjUvIPzg+iOogqrclquNt0faQUFQJU01OPFdTX8eubjj91XTr29nGdQuEl04WeOXYwTwx+LJ2y/0ab4GKvZQy5BtI8G53ddH2oOufW1pfGjnSEvb6FrshwoY6iwWyR0q5xdLm0hkTReyzZDA9GaCjRip+6ubpB/O1tBcwq/fhurjKMaB5COXZYIATwVFuPUA90loW8gjNSL/3U1sknIAnQKzCLBpetGPo8YnfMTFj3k3Prtle3LG9uGvoGWmL+By8xml6sgU6VrNf4bXsKq3rOeQyCMynqYKTL0as/WF0UDyyk8lKi0HlMRnicUY5zCRXUzXjSzdPMTeWKfnaqXT0xTDdrIs63HQ+4quHhjKdnLRZfcac0z5SugSOGBqMSUOgnXxbAWC+HNS7LRp2BHoJg1LjBMYRYyvnDSkea2xrWjDmJxH4DjXw/Wyd02gk95/N5/MRzngPTi8W/DOOiY4bj8c4aEfcZdAgU56fs9dJsrOYd8OG7O/FvvLYWmbzI7jKFUld8Tc8fqfZFpUunmCSY3Bjh+J861XnaUpl6Nhv7Igrgru4vL68v4QK/wazgwQF +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete a vector store. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/delete-an-agent-by-its-id-and-its-associated-sessions-and-turns.api.mdx b/versioned_docs/version-v0.2.23/api/delete-an-agent-by-its-id-and-its-associated-sessions-and-turns.api.mdx new file mode 100644 index 0000000..82adb66 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-an-agent-by-its-id-and-its-associated-sessions-and-turns.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-an-agent-by-its-id-and-its-associated-sessions-and-turns +title: "Delete an agent by its ID and its associated sessions and turns." +description: "Delete an agent by its ID and its associated sessions and turns." +sidebar_label: "Delete an agent by its ID and its associated sessions and turns." +hide_title: true +hide_table_of_contents: true +api: eJztV01vGzcQ/SuDPSWALTlGgjS6qbWLGnVgw1YOhWMUFHekZcJdbkmubEHQf+8b7sqSIrn1IRcD0kVckfP15r3hapF5DrWrAodssMhOT07kK+egvamjcVU2yK7+zJZH2ft9W6OCyfM/DYdIDyqQqWbKmpycp1LZifMl59lRpl0VuYpir+raGq3Evv8tiJNFFnTBpZJVnNcMt278jXWEYe1dzT6aNrsQVWzCxjkDt1P2OLid1h+j0TW1p0m7nCX/aKLlDdsQvammO6bn3iP5dPiIFIXC+UihKUvl5+QmFFExpzMPhdEFmbZob1QVCQWTqrr9FGcp/qMy9qWR29MS2roKtVHRlKo69qxyNbZMGzZb6UgkU6HmSr+gyjdXaaXsWxrSl5vLrhiN5MdMTeCcokNnYc0zptJ5RpnSz9Q5UmPXxBQ81KzNxGhyWjfeM8L/kBcSU3lu2njXGw2dKBv4KBP+GA+eDO5WHV516wm8+6f+tTA9A96KyzTxrkw5DK8venTjmmlh52iPte4h0M3vv9HHX04+9gQ0flRl3TJjxS9QfR3vV5XTTUvxdT4vZP5yKdW/P/20XznaGqiCCpgHWUTnYFzNV47FK4gwNTPGd+kanBFsTckHUR1E9cpEdfppHW8Epn8WpnfKCpvS+ss10AQQ4kfNnAtsIjYVmawpTezRtWWFhCLYo6YKIrHY9L1Obx+eu6kC+xmaDzhFSgx4hFhNxY/AO+KJOxgO0jpI6xVJ68PmfXUhzEbmdNuyfR2/k9dwl/IrMPIeXTVgAqsy3UpjxsVTuQjAsNfJK+eJamzcldh/OT6I6iCq1yWqk33RXiChpBKhmpoGqWY4BemDJL+d7Blbxo0GsJWcoPGcDN74Ls7wU56WKgSnjZIQ+HsWYBbSXmx8FXoiDOVVCSdeAi2yCg9wnNz9bURyRgLVKhY7UMl1iFBdk9oM0Ow8JSW+1w2JvkGDduTZkWq5RGVIonA4mrX2KTUEHWT92bt+ch76i1VeS2y3N3GbduMF0iLGetDv453guHABNR9bq0p1jHbo7z3tykwCBQbIJs6T4dlqEN3dy55QM+XXde1S7OlW7Om2I2giLuJL8BaI2bs92Iiewxaz1abWNj3HQkXChJoZOPla0Q8fDA9OL+1c5bXDqOpaWLDxmEq+JWsOJEnmmJ8ozaF1qjx/rYRwzicl7jofy58PC216YNtl9RmTzoZE6ho4YmwIJh01fgLntqBarIf6z/DdMSvyY+zXFu91UkWix6Ij1F3br5ZSWAyeyA4CCG/kyGIxxtvhF2+XS/kZL5heCINlGtFjYQfok5sg6/xpBj1b2pubTgxv6X91s7eG7keQW7inbCNPWH7n+aZgl/c4W2DI4zaTDNvtodZcxw3DndtzS4Jn55fno3Po8l+1wjZW +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete an agent by its ID and its associated sessions and turns. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/delete-an-agent-session-by-its-id-and-its-associated-turns.api.mdx b/versioned_docs/version-v0.2.23/api/delete-an-agent-session-by-its-id-and-its-associated-turns.api.mdx new file mode 100644 index 0000000..a1252db --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-an-agent-session-by-its-id-and-its-associated-turns.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-an-agent-session-by-its-id-and-its-associated-turns +title: "Delete an agent session by its ID and its associated turns." +description: "Delete an agent session by its ID and its associated turns." +sidebar_label: "Delete an agent session by its ID and its associated turns." +hide_title: true +hide_table_of_contents: true +api: eJztV11v2zYU/SuEnlogsdOgRVe/eUuGBUuRwHEfhjQYaOraYkuJGkk5MQT/955LybZcO4O77iWA8xLa5P0695xLuk4c+dIWnnwyqJPzszP+l5JXTpdB2yIZJDd/JsuT5O2+rXFGwtE/FfkgHqUXuphLo1NhncilmVqXU5qcJMoWgYrA9rIsjVaS7ftfPDupE68yyiWvwqIkuLWTL6QCDEtnS3JBN9n5IEPlO+c03M7I4eB2Wn+Mx7eiOS2UTYnzDzoY6tj64HQx2zG9dA7Jx8MnQgqfWReEr/JcuoWwUxFQMcUzj5lWmdBN0U7LIggULGTR7sc4S/YfpDaHRm5Oc2hjC9QmsiqXxakjmcqJIdGx2UqHI+kCNRfqgCpf3cSVNK/FUHwaXbfFKCQ/IVF5SkWw6CysaU4it45QJvczdk7Iia1CDO5LUnqqlbBKVc4Rwn+XFxKTaaqbeLedhk6l8XSSMH+0A08G96sOr7q1Bu9h3b8GpmfAW3FZTJ3NYw7D26ueGNlqlpkF2mOMffRi9Ptv4v0vZ+97DBo9ybxsmLHiF6i+iferTMWoofgmnwOZv1xy9W/PP+xXjjIaqhAZzD0vgrUwLhYrx+wVRJjpOeF/biucYWx1TkdRHUX1wkR1/mETbwymf2Smt8ryXWn9ZStoAgjRkyJKGTYWmwwkjM516IlbQxIJBbBHziREYrDpeq3e3j13U3lyczQfcLKUCPAwsaqCnoB3wCdqYThK6yitFyStd9376oqZjczFXcP2TfxWXsNdyq/ASHvipgITSObxVpoQLp7CBgCGvVZeKU1lZcKuxP7N8VFUR1G9LFGd7Yt2gISiSphqcua5muEMpPec/HayF2QINxrAlnwCl5P3jOpkITReflcX2ErjUnpvlZYcKlSu8D3WhHQyh73jGHVS4AM3uvHxt2a9aY5SypDt4MR3Ify3HVoFRq/TmBP73/QjuAr92VFnyylUuo4e6/ix2E3p68hb+YBoP5YJIAYkmcXRpPEXgUISg6Q/f9OPwXy/XuW57Leh+vUGuCVsmndCg2zluOFZCOWg38eL5TSzHp04NUbm8hRkUV97yuYJR/cECuiwiIYXqzF5/8B7LJyYdMupa7YXd2wv7lr5RFkhPgdv0Jq/2QMgTxu/pTvZnQRdzyGTQWB+zjWcfC7Ed38YbRR/UlCRlhaD1EfSwYt2mJmukVIKeAVPWTeVinzjVDr6XLAcrItzYtf5hH8aGUwOB8DbrD5iDhsfJVcCRww1xqTlz88pYgulenPb/KTblmSBnkK/NHhrcu6RFHXLrfumSw27sBh0dNDG4m872gQdmEVsWtcTvGQ/ObNc8td4DDumD5bxOpkwV0CmVHtep+t5+Wy1r0atXl6LA2S+t7r2S5CduShNxZ+w/EqL7RnD2v8fMjtgCPyHPNddWD7gbIarE28ExrLZHipFZegY7rxJtubJxeX15fgSQ+YbyWqsPA== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete an agent session by its ID and its associated turns. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/delete-an-open-ai-response-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/delete-an-open-ai-response-by-its-id.api.mdx new file mode 100644 index 0000000..892e52e --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/delete-an-open-ai-response-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: delete-an-open-ai-response-by-its-id +title: "Delete an OpenAI response by its ID." +description: "Delete an OpenAI response by its ID." +sidebar_label: "Delete an OpenAI response by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWF1P4zgU/StWnmak0gLa0ezw1l1YLRIjUCkPK0ArN7ltPOPYGdspVFX/+57rJG2gZbcj7QtSeSFtfH3PPfccf3SZOPKlNZ58crZMTo+P+V9GPnWqDMqa5CwZGnFdkhlenpOmQKMm4HryjdKQ9JLUmkAmcKAsS61SyYGDb56jl4lPcyokP4VFSZjPtoGlsyW5oOrcKuuM8cEpM8OYl1DujPpRkVAZ8qmpIifsVIScRBahZaKtJln12jw7JgVijxfr2mOeqaz01pfd5HXBgifrIOgJqZ/kwouHdeRDwukbSJ38E2s1SdPNFlxFr/NEmvEoAHOqXBHpFFMtZ+tcY4QlKySRWab4tdQ3HTKnUntM6+hHpRxDuGd2exvmW2yPvSSooBncv7b4JcD2tajna3GC27oPDBd9ka1u1l3pAzOj/mWXzMZoIyMmH8ST9EKZudQqE9aJQuqpdQVl/5/afJCh8p1xCtPOyG0V++d4fCPq0Sg0i8pqOPsvtV44B/BxMDonfG5dEL4qCukWrXApjnnKVZoLVRftlDRBoGBmsH4f80RNBan0vpnr0ZxaW4PaRF4V0hw5kpmcaDbNOuYFHM6k4BBp0j2q/HBd1gr8KIbibnTVFJMC/IRE5eHKYNFZRNOcRGEd7GO4n7Ww5cRWISb3JaXwVCpsmlbOEdK/wvUzim863HZrTV5H85GmN8hrNSumzhYRw/Dmsi9GtprleoH2aG2fvBj98bv4/Ovx5z6TRs+yKGtltPqC1Df5fpOZGNUS3+DZU/mNc06/7HZOqhVcIXKEe34I1iLYLNqJeVYIYabmhP+FrTCGuVUFHUx1MNU7M9Xpl02+MZT+lZXeOMt3rfWXreAJMETPKVHGtLHZZCChVaFCX9xgSwagAPXImYRJNF66dqf69NZO5cnN0XzQyVYi0MPCqgw9g28+iVBDw8FaB2u9I2t96u5Xl6xsIBe3tdo3+Rt7Dbcl35KR9cV1BSWQLOKuNCFsPMbGQ3PW2mt9Dt5x53hz4oOpDqZ6X6Y63pVtDwtFl7DU5MxzNcMZRO8Z/I4bI+24cInJQiic/S7P+yx+6WSBgY4nWyYGHzr33b/jFVHxfKUM+RYjvOtdnre9eJ0I3a0vlZxo04H6frvlx0ZFqxVKAaLcYmhzKY04kf4sGcxPBmiukYqf1r9TDJYdxCsMr7fiuqbKMad5COXZYIBDwVFuPXg90loW8gj9SL/3U1sknNgTWFZhEQPP25Xo/pHfsTYj3qZtVxwvbjle3DYKjcpFfk5eUzQ/2cEaG9q/kLbsmq07c8hlEFii5gqTPBjx6g+rB8VTO5mstFirPFoeTzTKYVlyNS8Z38N5IXNTmZKvJ5WOHgwrzrpoxe3JJ3z70DCnk7MW1VcsddpHVZfgEesGc9LoZk/RvaBjuVm5941vFBPoOQxKjQMao4ltXjZCua95r6XC/TiJEmx/1uolZ12Jo7msCQ5bLic4+t05vVrx1zg9OhYDHuP6O+HOQxqZ8vycrReYN0v6MGqE/1H8hFt2Vth8CQlzRVJX/AmP32nxyrOrRwzPsZxj32K49YhhmlIZOrFb++QL751fXF2ML2DIfwCjd/6V +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Delete an OpenAI response by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/describe-a-chat-completion-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/describe-a-chat-completion-by-its-id.api.mdx new file mode 100644 index 0000000..2e89ab9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/describe-a-chat-completion-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: describe-a-chat-completion-by-its-id +title: "Describe a chat completion by its ID." +description: "Describe a chat completion by its ID." +sidebar_label: "Describe a chat completion by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztHWtv2zjyrxD6ci2QOmlxi93Nt2zS3vmQtkGbYHFoioCWaJutXidSaYwg//1m+NDDkmPZ6W6dZooWtiVyXpwnOWBvg0KoPEuVUMHhbfDq4AA/IqHCQuZaZmlwGByx97lIj8bHWZLHAh/+KfV8nOalfiuU4jOhRsFeEGapFqnG+TzPYxlyHLr/RSGQ20CFc5Fw/KYXuQCw2eSLCDVMzIssF4WWlgQZNcYoXch0BmPaFJ3PBRufsGzKNHwL51yzsCIuuANa5pkMLTwHiRcFXwAgqUWi1lORWMa6wkDU7iWbFlliKEiySMQAI0vF+2lw+Gkd9CKLRR+XIEIFAgxKJQrD9JSXcetBk5S3pdJsItileX0ZMJ0xGcESyOkCyJKKcfjL8KUn2cimXqcOvY6Su73VchvKo327mkctbnSLx+pBP4/4egWP+Ip5toBDA2mQEjVnenWqRAWQeBRJHM7jswZzUx4rsQeW879SFgL09ZNF5TB/hk+pcYUDZzigoLXxHFt0Z7zQ5zAcPnnSJa5JWA5D2TQrnB2+QF0H45rEHd33xINBNtZwyxWSCUC6Kou4tUztp/1rVY1ZsWDmfXPFaqAdgxubsRcfTpnKRSinzq8wnkYM+AErV0A2i4TmMla4Bvcy7XDcrxqIzWmDJRWZSMO4jOBB2tYSnIuo10N99j63uvScnYprESMKO9csrsVUszRixzy18oyzb5fBHnyZy9kcv8Hwy4CXOrsMNtNTZL+jn0bEwPNWKl8v3XC9Nwj7FX/cVI0fpflTGYuW0psHIBzzuTZ44KiriGvexeFgXPVFOfcu5UkPcWuXpi36NwDojSF6iyU1XHZW04KDx5HEBUtkynVWIKWO+8U7Q7mHkkAWgJQjK22HTBFjRyNG2w9T9KDo8RSjx0AvT0Fk6yCC8AYutl3nOxN3Os643w/vsW9zGc5ZCAbg7c64bTT6DAZiORRJjrLql9M9todocZLH2ayujMLJUOY81aPNhGYqwro268jtAtC4Yrtf84+6VakhTaKvG2AASIxQekj2tK56VQsFxWJL9RuP+n2+HbCyhrWvv2MVS5nITmcig8zdaw1GvSQHAY/YeMoSWBEJIJd0BoRdmPh4DUsQ7SGAhXkEQCEnAIepRTQy5l2mkSjiBeYFpzFPOPuoefgVBoIfSTjMilXGQLrOlyzjeYbcihuOfO0Z1uFfwrVGgDrLYgjbU5matVHPR9/FC7Vp+Av90EeDaI0nWqbGCB25l2A+RRkazjELMWhAWcAwqm207+GBOKRASgP3LSNtP+231GrMKnP1ZP5DMb9zSu6I3NEKtXiwYVf6uMK2QT7gUa5CHscP2O2W4PFuGsMkMDbr2XFuEDrGKZ5K49WQBl/fxEC0qcwGbOY3gF6kErIQr5VSGNfZxmBYXpd5l6nxMe3su/mw3y78kJVpiB9QUVIB7RSeDa7eNGcNLjU3Vp13DbWpCAU+PK28mJUJsKQ2gnrkZyGoHNTRO+sKhZHMfz6+f8e2qzLaJn0OC30MFHuhbeE71jgNj6HrKCotbu8VDHMX1TnayFb1TcinYA5maTwGNWKvOZQoDcNRdZ7eTy+zirJNQO9I5Mi7lYE1Bbo6DjkLZjANV/esck/PK/5/RMWBYmwHQP9gRQCE1yuN3KxJ04lXHra34O7s/qz3YRZZLTCFQyiLeJpZRKUGHnhVvxsl3iZ7bynsPck8epaBDqAQORCKy+xcAPwCLcElchYj0+vMOcwf4AEi3BLFCS0Nbj/tV+NqzEqHUI2g/YenbapLzPbqxYMz/g7Uv7CaP/G4Nt1arInc3NYHnCA6uhsniKYBZq3dUF8N+Qbqq6GTUToZpb6azrvdPhKlvpqfOGJQXw1FD4oe1FdDfTVPua/Gd8E8uIylBhtKSajBhhpstm+wafTCPNgbUbMNuSZqtqFmG2q2oWYbarbZ/WYb2xzz4LBPPTeUTDzWZIJ6bpzf9O0xD/YG1H9DZkv9N4+i/6a7SdzCbja6U6nmV4Xg7u6e9cpox9b5CqSnWZ6LiM1EKgqu3Rb60NoHQcpm0WMv9EEQcTYDY52oe6sAnA8DcWdkwicylrhEjaTgq0jV8pnRuhqh4XG29CqItvdkYbLQgy4r8uK6q+XQeJuWyUQUNnvKr5py2m16N/JVhqYaWk98zU+z2Rm+63WaWb6kGAujFtwqhbWwgQZlcwhnUVtzsLRaPQzB+PtY+u7srHGWm9rVHfI/LRWPd18VyXTIdHbJdDbd3MEgeerlsZFs6yjYjv4+ZvckhiYid5MVG6q/j/z8Xo6/zw8GOIu7JztH+KPGfYDNRL3v3bJWWAwMwfsj728yjm0qvwQAOz0AMwhLi2hgYnORyhumJUhc8yTHNVcCaI8UIBNp35WG7BtUCh4J4LP3DQ5KzGw2phEeAikV5GRQh7i0TKy6QFHi7Y5XPiO/z1/RnYdUN+5M3Ui92dRd90S766itjnqzKWJQbzZFD4oe1Ju9e0GEerN3uTf7odUrtWRTJkIt2dSSTXcekjt6Qu6I2rCpDZvasKkNm9qw6c5D6r+mLIL6r+nOQzLVn8lUqeea7jykvpqn7huor4ZORp/oySgdiVJfDUUM6quh6EHRg/pqdi+IUF/NLvfV0J2HlJLsSkpCDTZPusGG7jwk1/SoXRM121CzDTXbULMNNdvQnYfUc0PJBPXc0J2HZLY/udlS/83fcefhRhxaZ+duJNqr7clf0OPv5+ncotNVtIrWP6Wej3H0Wz/4zhD1z4ODbupv3bnhzVzqAy6RxzLCzY+Ex7gxJJr+GObzPI9dtrn/xV3qqMK5SPh6xwDply77riZbXrN/n5+fMTva7GsFtdDX6ebrosDkBQfv4c7sPAMLUmWS8GLhlVSYMXY7XlqmC4mFq7m1K3XvDZ4NjsIsZjsaUcdZCryxeZnw9AUsaMRRsxpzWuTYq5IwQQ03s8Ajc45Yny1MRHU1UyFgNug6JMEFJrt2o88UQpOs1HZrztUPLAvDsigEoF+iaxOFdivsV6sSXkNfjZhWCK/KLCpTPTobj9iHrJzNY7xULY6zb4p9eHPMfv3t4FcTht1WZlO/QNVrfH/wiH2wKl7TM1DzneW8+r3fcsJYokOdw3RMNDCrgMnpwgM2t7BxNpPXAj6TrHTeVyaCjIqM6pEZ1avfa3yQh7O3qOnOslTTtP6blWATICFxEwoRodjQ2PCOuFgmUo/YWSw4EKRBe/gMinUWw8ti5Oztl1WRSoniGhYfxImmJEA8qFhlKm5A3hAvnXDJtMi0HpVp/dKMV2PUbKCcfbTaXuN35nXUVXkvjGjE3pegCYInJipNBASeNDO7OpE3r6ruWjax+wCTUZFRPS6jOujDNsCEjJWgqvGZQm7G6VQYiQSdqvrE/AJZ807BOFkwCfnf+GSEBoCloQCzVmafxRbZQT3abnpJBAkF6LzbP3ZSX17fQoPAa8nrooSV6NhhddQB9AMV8wyGBjNhLBOxHQb71y/3YUVTLvEb4tivcaj92xaldzDPBmLLjWk+DOZa54f7+5ASvJhnCqT6IsbehBcKexPwllvTo6sEyFjqhZl44v3Qp8/4DjXTUO0Wrdnb8LF5yAH4EbkVzfXLnt0dNGfVUmzeNLUmZHOdrWu7UJcpW/oDvkOYnF2kUZ5JPFvCdi2AIrE9oLC66poHYIGnHCp5C5QX4hJPIuKsMIbYBT7B2iMG0yxwd8JS9RbrfXsuk4McwWvUB2yD9a0lj8Z/NzAYgNMc3NPaz2NI0JAe12VqdeaTlbzVGlyRl2Yng5vdi1p34NdhW89hpVFBEMLt7QSywIsivrvDx5BIFqgZ8NW44gmqAehJJBV+jypfs5K9Zx+cLTxn95pML3t+rzPFnU6oBUv8BV+/ikXHWO8+w4Q5eHPc/AUS7ZijMBS5bszuhMmWCf7r9TkY5f8BIspWEQ== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Describe a chat completion by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/describe-an-agent-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/describe-an-agent-by-its-id.api.mdx new file mode 100644 index 0000000..1a072ed --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/describe-an-agent-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: describe-an-agent-by-its-id +title: "Describe an agent by its ID." +description: "Describe an agent by its ID." +sidebar_label: "Describe an agent by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztW21vGzcS/iuEvjQBZNkJWvRqXAs4SdO6bRLDTlEckkChdikta+5yQ3LtqIb/+80MyV1Ku7JsN9dr7lQEjaTlzuszw5khczUywta6ssKODq9Gjw8O8K9c2MzI2kldjQ5HRxU7WojKMT1nrhCM45fJaDzKdOXgI77B61rJjOMb+79bfO1qZLNClBw/uWUtgJCe/S4yBy/WRtfCOOmZEr2pzJOV1hlZLWDlqiS/VvJDI5jM4QU5l8KwuTadTKPrcSAGks3loq/KU/q9MSQos8I5YGPXqIy3yGt5CcpWi2nNDS/tdgVBG+7EYtmX5zUwjeRYXIam1ZV4NR8dvtlG2j/t2w0MYMExo4URIl+SIee8UWs/paK8aKxjM8HehgVvR8zpaOolmEfavqijazR5nkskwtVJItucKyvGAK8PjTQCnPvGC/kOzCudQml/IEZngehZpLkumV/WZw4ycQdOVGAYS/4r5KIQoAbYaMZnUkkHi/S5qBgsFDwr4F1RTwAnf9auTtfTesWs3S/DVqXntzQqmEiUIA13jUnFqJpyJkyPCaDaGa0sM7zKdVkJa5msWsIT9iPaxbALrhqBjzIjuBXJcmJJCgwyCzoeTL75qse7KRsF4XQhVq1eQF4ptMoptqomU6KxiUTPPE2L5kCyfw5Ir3V9shVGsGivZg+CLA834gl/BivNjS4JVLbkSiGqIF1QCkRAWXYpXcGyYe2/+5Z9MpSd91B2vgVl57dHGVHrBJCQzxcDAHtJUPDK19EAwALlBDY+DXfOjeKANRUAzbFH93BvlG7VzT/fys3nm5wLoAQbY7poF4Aa6OVzVmqQVMlzoULSsOBC4J5LJF/Kijtt0FjBh8uXvESpgrQlbIHoPVgQMuzWfWGXvO+XvHupapfEd0n8P5bEe1lyl8w/02SOckf2PdbXmMI/Tv3iDWaMIdQTE3jCy7JsSlYl5vWmRVkzXqERob/AhCByDG2UNNMIVCQzYUjFJ7xMN77dWurGICjLGvKiguDrRESSlYbM+DGDDEvESp0L9YVl1Jd9RH9VC1cQho2ogQvymdYCnOaWNyaJRxtAMxPuUoCAe48nBwwSD4O/J+xEW0nhE3ITcZB/CFaJy2iFGeSqnEHTdVkIh3kM/rdksGcKbqIxSGir2ZybcUxw5NVEN/KpLLTOydlcnTM+040LvGqZWVLYwudERW4MR3hJyMm2H6XXvU4TI4N9CekCGs4qA6VAbiNIlKOTY0gcSjHkEV2Kcs4b41XzgCOHGgHpv0IHoXL0GnoNXcSD2kQmMprcLbraBPCuD+0T356uqxYfM2pfhRPGW0xWdeOmtpBC5fYOpoNXwf73ftdprRZGN/WN7/X64dZx29JyRVXagMO5WQz27sO2HxChapRKBZiBJoJX6U8hspJf+nJ7XfuKXL+7vhMUSNGgVoIGGt28BtF+QCP/BhveEa64HlpCDsmUxBEK+uU2rryT2XtpBVbFsZIjAdZWbCPw4FXtjfOQ/diUvNqDnJHzmRIsWYccLjELRzYs14LKpS4E/kpNW64rIkw3FRBrew087RFiD8RkMRkz/w4mT9q0Ht7Znrc0Ysc4WrJDYi8ekn3FmUass/yt3RCgsukIw5dI08/n0G+yutB+yug1C2T/G8F5AxJDXey3w3a4mKg2pz0AYHQBBVd+t4wfwnwNNqvSJG+sVFxanbTQW5f/pHOpmMuKhCHZOZl+0t8jE41/kZbqlS6evDvJaTzLRO0IJfAMkOUGh8J/o7x7g6ZHrZQsKhMKkC6H3dmda04C/AxUwpi2Os801heRBipFWYpY89lJu6lOs0LLbDCjCLAPMueN0ylYxqMKrL0uzlNPZ2PcomQgzkrEIm6AeAmxmmEhBM/gD4cmHwrbUKVTSQc5RUJrX/IlxcQM40UpfSnyCTt2oDEUk7nFwhFfOYZWyjQZWeA5rcNCJuO1b9rAyDE1EvEJCV0bkWHN7dNPtI6vqqeAcBDyJiPRccZ4NG8q4jt1HJ/WS1fAFwXA78UYUX7uCffCzBfzni3FVwadKWkBjRZ0q/vsD2E0g77bkWntTUp05xxbetNVPGyqpv6fkDGQG96lzXmwwMZEMKRkq9i41XeMCqPtJuwp9ICQAMC3tciwl/eZlWEKQM2RCA8PZeaz7sqMpTP5BDmF4cQnx/JGlaONbeKkQmPbFfFM+iCgoft5smTBmmP2i+IlZ2eOZ+e+B+IOh2OuUzsSwAoDLDnDWQ1k1xp75RQVE7bH3q9H2QT1en8YOugoAuNGBLpIhSPKfjp79ZL5IBmmlJpmmCJ0aUqH7MvZP+ML37Zvoj+/g9Z0McwhsfYwA99NobwntJTZJTSLH9neHvBTYaeN3IK1qeVdWjDqtBTW8oWYzkTBL6QfWW8M8xqDiAK9VjwTN7rfH55Szgpe1xfCGChhyD3B2cyLEaYWZIIz+uWFl+tJEGvieYMJjmofyVQkhZooUgnKRAjclUdQC5ic+k92Mxk/DLxBCN/QrwmGgx0JiGiCFbyJ2RdXVy0eul3bXl9/QdO8KsfDcpEMFVp/JqtxE2hUjqlPVhayOaS+lQFicN/WgmMlZ/stY2D8nJyMtwU3BOckjsZkNYdyU651SwPzsUcH+AZG6y3amTaTJKf6ISv0Dvdll+QHpglDpIO30hf7ZG/XwcVA8Pl6hci4LcecUALqQrOk6VhUqWtZRIV91dQCevwszlgIaBz83Nw4BUdukGiu+CKCCtEXZ2yBDaMqteC0iQCYItfcN27+/keye2zgElfGXJ2lkPkEdxbSTWoaLo+MRwvoKEpu+pmpm5MnLwxNy+PW4vukOI5f5dLFVPo7mCf9+nk1LQh/2u689ITW1oUhs6ALwTpgM6ipcDsLO87Zs5/H3ngS6yaHh3Dsfb3MOdg1e9/WUPc4CEgNmpSTP8HPZ/TraZBxQwHdT1SJknuLhhJ3q2eczurqz5zH3xuZYcntURlprp75+t/A2LNq/hmi8MnL5ywo0Za33n03Y/Je+EITrRxTE987gyrIezOg7nU7YEtG2eXKXa78O+TKFkx/PUZ3WfN/Omumx/BrHO4kuu8w1rqC9WOt0O7guZYROECbDk9IYvU7glpZ7OFItd9PwI/W8bLG+rrqqn92CZV1oH43Ddqbz2v3lleEXdeoJ9dRFeRAQ3BoJcK1lhXvYDsSB9bkA5Dzy6Hb3v7Q+kODAxhUTFYXXEma25VcoZloHPiJ7n6DwK7Z1E6uHEy9fn3C/GrQLBejDkTbGrfvjcFmDRePccBWaAOtf4NAXcahoKA1fuIovdJGQgL3JyBVeE58KFgdl7dobj1nvxpZK12Bbqy48ZCtFSd2vLwanuJvPqNgv57+EpQJF0+oP4VsaqA3leICe26DUwUPewJIe4jRTiB1ljXGYHO6JtfdbiiQh6O3WuMlsCYzbTBe13TGq1xHJ8cTdqqbRaGWYSRs2enzp+zrfxx8TTuX+EjXv1J8AdQ7fk94zk49xDt5bon8EDmPvxmOHH94Tr22FTTQ1vBytYyErZ8aLuQFlgllvOcT8s0uqHZB9TkF1eNvVmaL7AUiPUSWTUPrX7qBmAAL+dtq4b4aXl1iCponN2Enim6z0uRsgTejFDw0caf6atNOZYW5AOeDOTGUBB4wgW+aSnwEe+PUXwQz7EJrF1qfUWh9le5Xx4hsnMCeebR3/EN4HfUhH42RT9irBpAgeEm70gyvb1aaBu15DK/kIk2vtNxIeBdUu6D6vILqYIjbLUKIogShxvHG6Bvfhtn+BOMZfcO78rEjmy2ZhJrv+NlktHrZ8E28N5j2gBKJ1NwVPTMcP+v949vOvP5SXS/Y2ju27+jeVaFznIoICj/kcTjav3i0T/Ts/lUU4xoe+03VS9kYtE7hXH24vw/b+16h8eRoT+GJ/p7FE/1JpkuahVoB9pJ4tRxefBZzypt3+AxRRsIFB6Q3As7S6QbwR+Ze74tHA6d6ftKWgpSnYZNSpusE4TjXvq3Y2n88/nMQUeW1hqxjqU8GKtJAgjEedzkeqWFKMvNwggxEuRFv8Qq30oaCqk+crjAoCDOD58Reqhc4svCn9TXYETJAd4N0G3zW7nG2uXfbewEOeDtsv1ZQWiF3cutVAMIbb2cPBfhw2GISHIf+xiVXV3iF/1ejrq/xZ6jxDDoaPlKWnKFXwe25tPg5Hz6yTMV+cBoQ/JANwHtQ6jhIq/B6Lt2uhG/w8Vws00iigVoBmRW2EJTJPz6ia4jJi70tayVSfvj+NcTOvwHKkUx0 +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Describe an agent by its ID. + + + + + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"input_shields":{"type":"array","items":{"type":"string"}},"output_shields":{"type":"array","items":{"type":"string"}},"toolgroups":{"type":"array","items":{"oneOf":[{"type":"string"},{"type":"object","properties":{"name":{"type":"string"},"args":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}}},"additionalProperties":false,"required":["name","args"],"title":"AgentToolGroupWithArgs"}],"title":"AgentTool"}},"client_tools":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the tool"},"description":{"type":"string","description":"(Optional) Human-readable description of what the tool does"},"parameters":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the parameter"},"parameter_type":{"type":"string","description":"Type of the parameter (e.g., string, integer)"},"description":{"type":"string","description":"Human-readable description of what the parameter does"},"required":{"type":"boolean","default":true,"description":"Whether this parameter is required for tool invocation"},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}],"description":"(Optional) Default value for the parameter if not provided"}},"additionalProperties":false,"required":["name","parameter_type","description","required"],"title":"ToolParameter","description":"Parameter definition for a tool."},"description":"(Optional) List of parameters this tool accepts"},"metadata":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"(Optional) Additional metadata about the tool"}},"additionalProperties":false,"required":["name"],"title":"ToolDef","description":"Tool definition used in runtime contexts."}},"tool_choice":{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model.","deprecated":true},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"title":"ToolPromptFormat","description":"Prompt format for calling custom / zero shot tools.","deprecated":true},"tool_config":{"type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."},"max_infer_iters":{"type":"integer","default":10},"model":{"type":"string","description":"The model identifier to use for the agent"},"instructions":{"type":"string","description":"The system instructions for the agent"},"name":{"type":"string","description":"Optional name for the agent, used in telemetry and identification"},"enable_session_persistence":{"type":"boolean","default":false,"description":"Optional flag indicating whether session data has to be persisted"},"response_format":{"description":"Optional response format configuration","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"}},"additionalProperties":false,"required":["model","instructions"],"title":"AgentConfig"},"created_at":{"type":"string","format":"date-time","description":"Timestamp when the agent was created"}},"additionalProperties":false,"required":["agent_id","agent_config","created_at"],"title":"Agent","description":"An agent instance with configuration and metadata."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/eval.tag.mdx b/versioned_docs/version-v0.2.23/api/eval.tag.mdx new file mode 100644 index 0000000..b754223 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/eval.tag.mdx @@ -0,0 +1,19 @@ +--- +id: eval +title: "Llama Stack Evaluation API for running evaluations on model and agent candidates." +description: "Llama Stack Evaluation API for running evaluations on model and agent candidates." +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/evaluate-a-list-of-rows-on-a-benchmark.api.mdx b/versioned_docs/version-v0.2.23/api/evaluate-a-list-of-rows-on-a-benchmark.api.mdx new file mode 100644 index 0000000..e36bd27 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/evaluate-a-list-of-rows-on-a-benchmark.api.mdx @@ -0,0 +1,68 @@ +--- +id: evaluate-a-list-of-rows-on-a-benchmark +title: "Evaluate a list of rows on a benchmark." +description: "Evaluate a list of rows on a benchmark." +sidebar_label: "Evaluate a list of rows on a benchmark." +hide_title: true +hide_table_of_contents: true +api: eJztHdluJLfxV4h58RoYjbSGncSbA9BesZw9BGmNIPAKY6qbmqHVl/uQNBHm31MH2c0+5pJkQ9rQD/aom10s1sWqYrF8O8pVkaVJoYrRi9vRNwcH+J9QFUGus1KnyejF6M2VjCpZqhMzUqTnv6qgFEGalFInOpmJmUpULnF8IWQSiiJIAe5kNB7hIJWUCFVmWaQDGrX/a4Ggb0dFMFexxF/lIlMwGcOGD7M8zVReakbMmcAZLPNcLmCsLlVcDAGRYajxIxkdt8Clifp4MXrxc/1FUkXRaDmu/z5P00jJxH2UVPG5yt0nRZnD4t0njJDzwGCyPFsuxx2yfpqrFt0u8jQWJTxUTG94OAFQI6blLqtbT0qCN83T6y+KkrgqlEQgVRWV4iLNhZLBXMA6J+KN+SU0iKeIZSbSCxDfqIoTkchYiTIVSHRFBJezWa5mIPDhlKE9doJ06fGeFxgrABLUC2yWxWsdISGHV3Eho0KNwTT8VulchbAaV2oGKXQ2HpW6jBCxU2bFCb0ZdZE7HGKVFAU8iRSxa7SSwWqVluyyEteU1NrloN81d70VIDbWajI+Mumggwh9O2RL+dvfKlWU4loWQifwnQ4F0CCWEZAiBiQfzGoWpSwrV3o1gJ2BpHWX9MOnT8eCR4NehApZYOjRFcnup2/yHJCnwWPk4zzNS1FUcSzzBUoh8YrGXM81qKHmRedaJob3iXlP8xDvYVuJtp2ZR+PUUZrA2sS8imWylysZynOQKOebFjo4k05gzUmwxSqffcxYuL4Wh+Knk3dmMQEgf65EVYBSgYrlqHHqSokYhAqWifwkzgl5nlYlTV5kKtAXoJdpEFR5rmD6Dl47qSVz2HKrJp4rzwR1mHhtMUYcDo+PJuIkrWbzaAHsiSLQeHHy9pX4818O/kyqqW5knLFkWPkCUW/meylDccIi3uCzpeQbzfnm+2HNCSINWiHm8HmBP8o0hY+ThQWMUEEQZvpKwX/jtIIxSFsdK69UXqmemFJ9830z3yeQ9Pco6UazCle1/pNWoBNAIXUTKBUi2VDZYA8TkY51ORHH4HEAQiVIj5yBzy4ieJnbneq7VTtVofIrYD6QE1VJAXlQsKpE3QC90ZFQhgxetbxqPSHV+s7dr45QsgFzccrS3sxv1OuwL/KWGOFEfKxAEpSMaVc6V7DxJGkJBIN3Rr1CdSHRE+6p2DrAXqm8Uj0tpToYmm0LFSItQVGTswJXg/EXoj6chEIW64LcOgxGBZIVqJ8Ec5C1S8w4ZTKHiBd0uqAoHMNfjLrtkKlG3dIIM5PlfDC6O3ptuVF/Rpytkk7oKTDcc5lR5pUa91XThvPLMx4MO/jLNFzgiAfScp1kVfnFZXWIxUB5Q3FVJ8QA7vSiSoJtM4I1KmtSRzU8nLFCfUB71E/INaIEvLvQsxWxCr2rONNQg2pJ6nqO4qxTsC2hDmHlKyaxr9tEGvcZuGISfts3fIA9hG8vRjHsA9HI2cPMEyAD/9hoNBFNGopa1WMlWhHkJSntNsnLEt3a2WKFv2rACTvsIUkxy8G3XrRo4Txq5d8qMFCwI3w2Az6PcOE6BDXXFwuQA9j+eqjuZvIJSce+/5MmOjVATy3MLmY8rD854CRhw1YREKYgQZ3r2RxDdaDRuTzXkS5hUHqJcXXJmdWiVNlkQJd3pWuZZtOsRdbmyTBV6f2WRAUSqTjDhF+Vu2gYa9Wd5BVY4zyNCoiekjCNE1VQTsECnogfkC45Z1DxVZBTZNUMpylpAYOTmTUeTL7/rjd3FVcQl2lwHVpUn8MGPU+jkIxIUgWRqgoHo9cMk6wWgr2fIH1Ks+ONYgSD9jLxzODy9Up5Ih/BydkWsYwilKpCcWIGBaoQ17oEL2p49f/4u3gwKbvsSdnlBim73F7KCNpmJ/sDiQIvPrMEgCkQT5gmJyY3zLXoADUxhC/F8zuw12LXZvO/tmLz5SrmglDi8ULp0AOWgVy+BHsPmEb6UkXGaBTAQnTpNIKPdSJL8A+BWIaHiw/spBlsY3CEkHt47scWduO+4I333Yx3z1R5I+6N+O9mxHtW0hvzJ2rMl01uaNSbGoMDeTPlwSvIaFVoMFyQNzquYpE45GXSIq4m3WKOcVWIql1SwIWCSlGaQChs8AJ7CLRIqxyFMs7ALkagfA2KCDJJS5M+J2AUsXxVUI2LukF+JbNyTjKcqwxmwXmmmQKmlYu1RuL5CqE5V+U1Zgv3vpkcUM0M/HcijtNCk/oY20Qz6P8qkahrS4VziZkmCCqv56pEOwb/WgjYM5XMLTEI6SIVFzIfWwNHXHXWRjzV8zSltFUpo0uToeK5Mh0UHKTB73uE2D+hZohvwVz8VmGiq0C8c2UTTWA4okjgHHU1DMXiOS+NBU7w6SGY/wQZhIujz5BrpgyJbROCsRPtWBdQG4Czvmgfc3jaXZp9LZqcE1NsAftrPI1hJ5EztdnewWa11t4xvJbBcx4NWzweMGTyJBbB8GthUVy20s2D2Qx8adNiFrxRqM+jiTi6EGD7Sw062AEO0+W0H1wBHuHYiGtOQAMgeYJazCyuErCa0QJJ+i6SsRSnpQwuKRsN+gpfRSDTwFGRknR053lGZT+cmhyT8eXka8kGLY0EEFAnJA/F14PJASdn9VCJAx0jhV3mNU9cOr92vWNKhFtyW+qjyk3EYXQtFwXwgMB8HlESO95K0Ko8GjiGoFy2mYoA4aF8KEtJL4xqMSVxGL55QeP2b3G65V/RIv3p2/EtvllOxIcUE1Joq/F7sMDg9tD5JFtRPpQsNmfAqlxvl1/Cafi1yFLYZZz9C3Y2MPzBjtVOOLNjBgA+2TVY3gpGI3veJCCnFLeMmCA9XPkxnafiKS3TmgiNFQ02R7wBzzbII4JBKt2BbgjChyu0gjv4GixXDilovle84KMhC3RoluXK7EPE72Dz2x6ffXAvFUIorEEEb7O04dZjLeWdnDecxvXd4O/15CydKQ017xRKb2kivDXz1sxbsw3WrGstvFG7v1FzQ0qqyYAY+UqFLpzeSaIThXhvzds3b9+8t/b4DJv31rw189bMe2tflFHbxltbnq0bsxvulCJtkpRujpaSgO9NInPgulkrSWjykCidWPiZV6bkC3hlc/1GeylNvmMO2dpxUzXVLW9ysH6PQ17V1V19tLlmqinwotRmqxDtvvIM1EjaAs1POBe8Y2Ebfdqgu9nA+tKvp1894KsGfNWAL/3y1QK+9OsJGm9f+uWNuC/98sbcl3750i9f+vVQpV98D6+YaxWFu1xMg0+B/nf+FoubZnlaZWu/W1fttN4s8y3KAYbLfPbobxbuIgq0ULMsRxoOMb3xCVD7JxL537DhHeKI5dAQYgi3KpkiX3a6kbkN2XtmBTs8mSRjSQh0RmwC4Fxs/mHtxelrtMJ2GhGmitwl98btH7fSetYWCtNVDkRnr3HysvW34pmazCZjk71G40mb1tc703NLIjYTW0o2ktjTB2df4UvG7Sn/XW8I4Nk0gOEPC5NTdcg3nVylfM2YV1a3J/jjlXONJBq/mLfDOs/oLO2C9gBbV7qbxTdq3hGbNjbOFy2PK42Oa9Hr4n/csLSuNDUdzpD0A73NnBW/MxfbG31idhLTZBCorCQpgXeycx70GO3umpUe1lgKuxinO0Jpjehu7OwwCeRnwBNu1QBzzwZwIHLwFHWsrM9XTOpNdRrMUz3cHEIBfXByWZWpKyzjUQLU7qLziuGs1FvEDC+duxqLcgPAsX1EgI4Qts3A07a5TtrHFKYTBZZIo06cK9O1Aeuqj0pYMTiTITVLwE+OmoMP8ZbGoSMTyIyDNiCyNY3mDASRznKFxdqmx4GlDnvVUz6gXUck6mcwHtk79tNS4ttsUc7hD+zo0NMxgvyWAffUjJ15cy6M+hVAZEqrgEALotV98V+Vp9iIhLqERcW6RTSnHRti07Y8rPKm/p8kY8A2nLVOlJgCKw3B0CLrhY3r9Y5xwUi7iXgFMSBdA+AeKgtjWes2lwhENg1WyOq2ciwNySc4k0lOPLgsr1yypXHhMGmeYthl5ZnWgwIN0c/LhTDUHLeuRFAMJEtMjpXNsi0A9DA0tiOCDQWsa1Zya5qGcWJP/NLVsgmu65cXJoK2KNBNDXONAvt/oZT9ePrxg+kCPAzJJc0wRIjSotRYXyn+Zj/4e/0l8vMfEJrOhmdwqD08AUdTiO8xDRXFAoLFG7G357SQsbMZavfv7kzP1VxeaU5Zr1TzDJWIFD2LZKDWsv8VGRuyWYbr6ZXKc3BhiD2G2aJ1v4ZI0DrnfmnQmvDcQILDjDWZnCTjE3XPvo0I7DqHWRZMcsK/itVgOBm4BgkO6DuIYWJHg0RUhgqmfuSr29taHpybO8vlV5TNS0LslqOcpELNT2c0bgJYBXSOzZgKsObYmatlo5h9Gx2Ols3mLWMg/dw9GzeGbWJTYzq5AHdTd6KlgfzY84PlHZqsmBSnJpva6mJTH+67dQ/bgTbcahVM9MBuF8FZRWB73QIyrt2xUkUKuxcvKDtml9SELCrBuGpagPRwLi4vQKFVu4nXQOA0WMVUY3QRyZkVKpQ+m2Mz03CxFHZ5A8KCMNlZQw7cuIeWs3usmKXptsW2ulVO8QA1C+4mNTXdo8ajGUQUscz7lqnJkzsfDGXL7dbCcZJNx7dnaXTKfQ7kcf98WkELij9td4y9retjFhrLgiwE6gDNwKfC7czsOKev/zVm4mn0m0o8hBO/ZItQAl2DX+5TYeQS1HEnf4THp/TUdq9e4UD3DZWzyL1ZRYa7XmfTMvs+5/F3lkwzZHuptDDbZ778DIh9nlw8QSl8+eGtMIuo3Vtm33qZvJN8IYlax9Q0785CZfBdL1B3qg7YYFG8rfS28jHYylqY/ngZ9Vbzi7aa7jF8Z4adULe1yq2ooHusZcKdu9DEJNd6EFeXPifdYuJ+7fOdtoxuLOVbYvrSOl8X7UvqnmpJnS+le+qldL4u+hEab18X7Y24r4v2xtzXRfu6aF8X7Vti+paYviWmb0vi25JYX8M3WXqM/UjuFEr7Jkvemnlr5pssPVaj5ltievvm7dvjsG/esHlvzVszb82+DGvmvTXfEvMLbIlpOljeW659a0xfReBLwHz1wGOvHvBVA0+9asCXgD1C4+1LwLwR9yVg3pj7EjBfAuZLwHxrzLVm2bfG9K0xfWvMbenpW2P61pi+NeaalfrWmD29fcwNEBFp3xrTt8b0rTF9a0zfGrOv5r41pm+NuaLNjW+NuaJFpG+N6VtjPoGgxbd7860xH4cUfrFN3u5UHeBbYz5OKfW20rfGfEzy+MVazf/r1pju4t/AiwYaRvBBmq8vLN5aZjfoqtWcKIqnspj+WoWdSze9F13xLM05lMG5CWCbM4CxkHxloANslaHoDMshrrxBWhQUXp7LQgetc26e+W3CB914MEa7EH4+3TL8PGqCTpMefPfuvQlIbW4K80YElCtjeFY6uuO0G2ayIuTgLqeVrzjba1IPFkQdVvJ8dedSXhNOraZEFrXNgeWq4oYThkC9TW/KXAZ8eCGT4hrLf7BmqilUsTrM5fszmJ1EeWr5vQMiTt7pCqATk6+Vns1hnmnzKFahpugXUyQzoHYgMdNdUcm/DAIwOMGirZY1Wm8NVp/6h2J8dEppYGcdtdwWLndx1Zh1HTj1Ohz8FkgJHh4VG3EaCFlFc1HlYp5e380BcGR5WApWMcUhD0j0YfEjftvRmNVngEwM+HBPFnssiz09b4X/u0RX1vZ09LuxPb0X97E9HWC/r+3BD9CAP5COeg199Bra5fgW+kisPSbR2lEhaRLBUvk7KCRLuauJzZP7qKCF8vvqnhf9P1j0Nwr6S+TcjiJO3N4s23fKvLTkbOuw1rvK3lX2G/Gjt0ZPylVuK+zOpsh7zt5zfvIK+5g9ZyO2Oyum96C9CjwRD7p1Ga4zQ48q72VW3/3qQdYhFbtkbTT4xm537CKtxLXkSs+8opIYkJmp6YC8qq5ohWOIWtRct7MwWv/LJmzy7Nbao0ZFzVBzAwvrY/BiKJUmnqv68x1L8/GzaeAcD3Ty+S5jVRLMY5lf3uV8gm9qgdAVzhyN/IDCW+jT/hnGG7O2E/j8BC+dFSULQzODqTAGZs5T+HOUpQWZPlnO4a/9q+cyyuZyH5e7X09V7N820+pwud8cexCaKr+iwrGfTV+t0bwssxf7+zJZ7M1TrEbai7BKdK/AKtFJkHIfskKBQdB4XRE+fG3N6s9n+A4PFs0SXqbh4g1zFWXIoRB8R3B6ZOLbc2i4zgYoBkA67GxE0zqZJnBq7N9AZxi394v53NzhX7YvnB4MX9J8bq8yNtj2Sj0RNjdkb5qs173R3bt2vfOlnvodLKnM7iIlfK1X7FTvnronkTAPMpUV8+r5wDbHp+JkMltHmDaCdCBT6a/R0+JzIjr/SHt1WyUhtS0ryOoAFI2NkHL25UybJLBCF6baE4DKXH3G65ZRmlM1cR84lRtHivcogxX1O+ItAhUglklz26vWIqciF6UNq9GlqIVp0qXIrcOX7UEYucH7HfsQ+2oym6Y5HSvlzyOjlLgJA1zXBqD2vXBV04yxugnSj/qHQG5v8ZruT3m0XOJj0KwcFQ9+XslcY+UiqWGoC/wdDpclumt8dmKMChvro9eW8TU+ZiOghw1WAreoFQu3zkiCrghdsYK/4OelWqC/5S6UTtbnSoYYZAHiPMR089ozToQFgRu8Ec99qpPHBCp/cUg3mNaOPXPs5fHH00/IgZSbgKCZwNhMXqNrBv8mdNOsdq/o2e0oksmsInXmOv0l/vM/erY1KQ== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Evaluate a list of rows on a benchmark. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"system_message":{"type":"object","properties":{"role":{"type":"string","const":"system","default":"system","description":"Must be \"system\" to identify this as a system message"},"content":{"description":"The content of the \"system prompt\". If multiple system messages are provided, they are concatenated. The underlying Llama Stack code may also add other system messages (for example, for formatting tool definitions).","oneOf":[{"type":"string"},{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"},{"type":"array","items":{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"}}],"title":"InterleavedContent"}},"additionalProperties":false,"required":["role","content"],"title":"SystemMessage","description":"A system message providing instructions or context to the model."}},"additionalProperties":false,"required":["type","model","sampling_params"],"title":"ModelCandidate","description":"A model candidate for evaluation."},{"type":"object","properties":{"type":{"type":"string","const":"agent","default":"agent"},"config":{"description":"The configuration for the agent candidate.","type":"object","properties":{"sampling_params":{"type":"object","properties":{"strategy":{"description":"The sampling strategy.","oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"input_shields":{"type":"array","items":{"type":"string"}},"output_shields":{"type":"array","items":{"type":"string"}},"toolgroups":{"type":"array","items":{"oneOf":[{"type":"string"},{"type":"object","properties":{"name":{"type":"string"},"args":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}}},"additionalProperties":false,"required":["name","args"],"title":"AgentToolGroupWithArgs"}],"title":"AgentTool"}},"client_tools":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the tool"},"description":{"type":"string","description":"(Optional) Human-readable description of what the tool does"},"parameters":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the parameter"},"parameter_type":{"type":"string","description":"Type of the parameter (e.g., string, integer)"},"description":{"type":"string","description":"Human-readable description of what the parameter does"},"required":{"type":"boolean","default":true,"description":"Whether this parameter is required for tool invocation"},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}],"description":"(Optional) Default value for the parameter if not provided"}},"additionalProperties":false,"required":["name","parameter_type","description","required"],"title":"ToolParameter","description":"Parameter definition for a tool."},"description":"(Optional) List of parameters this tool accepts"},"metadata":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"(Optional) Additional metadata about the tool"}},"additionalProperties":false,"required":["name"],"title":"ToolDef","description":"Tool definition used in runtime contexts."}},"tool_choice":{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model.","deprecated":true},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"title":"ToolPromptFormat","description":"Prompt format for calling custom / zero shot tools.","deprecated":true},"tool_config":{"type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."},"max_infer_iters":{"type":"integer","default":10},"model":{"type":"string","description":"The model identifier to use for the agent"},"instructions":{"type":"string","description":"The system instructions for the agent"},"name":{"type":"string","description":"Optional name for the agent, used in telemetry and identification"},"enable_session_persistence":{"type":"boolean","default":false,"description":"Optional flag indicating whether session data has to be persisted"},"response_format":{"description":"Optional response format configuration","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"}},"additionalProperties":false,"required":["model","instructions"],"title":"AgentConfig"}},"additionalProperties":false,"required":["type","config"],"title":"AgentCandidate","description":"An agent candidate for evaluation."}],"discriminator":{"propertyName":"type","mapping":{"model":{"type":"object","properties":{"type":{"type":"string","const":"model","default":"model"},"model":{"type":"string","description":"The model ID to evaluate."},"sampling_params":{"type":"object","properties":{"strategy":{"description":"The sampling strategy.","oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"system_message":{"type":"object","properties":{"role":{"type":"string","const":"system","default":"system","description":"Must be \"system\" to identify this as a system message"},"content":{"description":"The content of the \"system prompt\". If multiple system messages are provided, they are concatenated. The underlying Llama Stack code may also add other system messages (for example, for formatting tool definitions).","oneOf":[{"type":"string"},{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"},{"type":"array","items":{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"}}],"title":"InterleavedContent"}},"additionalProperties":false,"required":["role","content"],"title":"SystemMessage","description":"A system message providing instructions or context to the model."}},"additionalProperties":false,"required":["type","model","sampling_params"],"title":"ModelCandidate","description":"A model candidate for evaluation."},"agent":{"type":"object","properties":{"type":{"type":"string","const":"agent","default":"agent"},"config":{"description":"The configuration for the agent candidate.","type":"object","properties":{"sampling_params":{"type":"object","properties":{"strategy":{"description":"The sampling strategy.","oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"input_shields":{"type":"array","items":{"type":"string"}},"output_shields":{"type":"array","items":{"type":"string"}},"toolgroups":{"type":"array","items":{"oneOf":[{"type":"string"},{"type":"object","properties":{"name":{"type":"string"},"args":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}}},"additionalProperties":false,"required":["name","args"],"title":"AgentToolGroupWithArgs"}],"title":"AgentTool"}},"client_tools":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the tool"},"description":{"type":"string","description":"(Optional) Human-readable description of what the tool does"},"parameters":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the parameter"},"parameter_type":{"type":"string","description":"Type of the parameter (e.g., string, integer)"},"description":{"type":"string","description":"Human-readable description of what the parameter does"},"required":{"type":"boolean","default":true,"description":"Whether this parameter is required for tool invocation"},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}],"description":"(Optional) Default value for the parameter if not provided"}},"additionalProperties":false,"required":["name","parameter_type","description","required"],"title":"ToolParameter","description":"Parameter definition for a tool."},"description":"(Optional) List of parameters this tool accepts"},"metadata":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"(Optional) Additional metadata about the tool"}},"additionalProperties":false,"required":["name"],"title":"ToolDef","description":"Tool definition used in runtime contexts."}},"tool_choice":{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model.","deprecated":true},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"title":"ToolPromptFormat","description":"Prompt format for calling custom / zero shot tools.","deprecated":true},"tool_config":{"type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."},"max_infer_iters":{"type":"integer","default":10},"model":{"type":"string","description":"The model identifier to use for the agent"},"instructions":{"type":"string","description":"The system instructions for the agent"},"name":{"type":"string","description":"Optional name for the agent, used in telemetry and identification"},"enable_session_persistence":{"type":"boolean","default":false,"description":"Optional flag indicating whether session data has to be persisted"},"response_format":{"description":"Optional response format configuration","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"}},"additionalProperties":false,"required":["model","instructions"],"title":"AgentConfig"}},"additionalProperties":false,"required":["type","config"],"title":"AgentCandidate","description":"An agent candidate for evaluation."}}},"title":"EvalCandidate"},"scoring_params":{"type":"object","additionalProperties":{"oneOf":[{"type":"object","properties":{"type":{"const":"llm_as_judge","default":"llm_as_judge","description":"The type of scoring function parameters, always llm_as_judge","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"judge_model":{"type":"string","description":"Identifier of the LLM model to use as a judge for scoring"},"prompt_template":{"type":"string","description":"(Optional) Custom prompt template for the judge model"},"judge_score_regexes":{"type":"array","items":{"type":"string"},"description":"Regexes to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","judge_model","judge_score_regexes","aggregation_functions"],"title":"LLMAsJudgeScoringFnParams","description":"Parameters for LLM-as-judge scoring function configuration."},{"type":"object","properties":{"type":{"const":"regex_parser","default":"regex_parser","description":"The type of scoring function parameters, always regex_parser","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"parsing_regexes":{"type":"array","items":{"type":"string"},"description":"Regex to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","parsing_regexes","aggregation_functions"],"title":"RegexParserScoringFnParams","description":"Parameters for regex parser scoring function configuration."},{"type":"object","properties":{"type":{"const":"basic","default":"basic","description":"The type of scoring function parameters, always basic","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","aggregation_functions"],"title":"BasicScoringFnParams","description":"Parameters for basic scoring function configuration."}],"discriminator":{"propertyName":"type","mapping":{"llm_as_judge":{"type":"object","properties":{"type":{"const":"llm_as_judge","default":"llm_as_judge","description":"The type of scoring function parameters, always llm_as_judge","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"judge_model":{"type":"string","description":"Identifier of the LLM model to use as a judge for scoring"},"prompt_template":{"type":"string","description":"(Optional) Custom prompt template for the judge model"},"judge_score_regexes":{"type":"array","items":{"type":"string"},"description":"Regexes to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","judge_model","judge_score_regexes","aggregation_functions"],"title":"LLMAsJudgeScoringFnParams","description":"Parameters for LLM-as-judge scoring function configuration."},"regex_parser":{"type":"object","properties":{"type":{"const":"regex_parser","default":"regex_parser","description":"The type of scoring function parameters, always regex_parser","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"parsing_regexes":{"type":"array","items":{"type":"string"},"description":"Regex to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","parsing_regexes","aggregation_functions"],"title":"RegexParserScoringFnParams","description":"Parameters for regex parser scoring function configuration."},"basic":{"type":"object","properties":{"type":{"const":"basic","default":"basic","description":"The type of scoring function parameters, always basic","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","aggregation_functions"],"title":"BasicScoringFnParams","description":"Parameters for basic scoring function configuration."}}},"title":"ScoringFnParams"},"description":"Map between scoring function id and parameters for each scoring function you want to run"},"num_examples":{"type":"integer","description":"(Optional) The number of examples to evaluate. If not provided, all examples in the dataset will be evaluated"}},"additionalProperties":false,"required":["eval_candidate","scoring_params"],"title":"BenchmarkConfig"}},"additionalProperties":false,"required":["input_rows","scoring_functions","benchmark_config"],"title":"EvaluateRowsRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/files.tag.mdx b/versioned_docs/version-v0.2.23/api/files.tag.mdx new file mode 100644 index 0000000..bd62d0b --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/files.tag.mdx @@ -0,0 +1,19 @@ +--- +id: files +title: "Files" +description: "Files" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/generate-a-chat-completion-for-the-given-messages-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-a-chat-completion-for-the-given-messages-using-the-specified-model.api.mdx new file mode 100644 index 0000000..b29fe9f --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-a-chat-completion-for-the-given-messages-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-a-chat-completion-for-the-given-messages-using-the-specified-model +title: "Generate a chat completion for the given messages using the specified model." +description: "Generate a chat completion for the given messages using the specified model." +sidebar_label: "Generate a chat completion for the given messages using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztXXtvIzeS/yqE/skMYMuTILm9eC8LzDOZ3UwyGDs4HCaGhmpREjP9SrPbHp3h775VRbKb/ZAl2ZNE1lYQJFY3u0gWq34sVhXJ61GhTJ6lRpnR6fXoqydP8H8zZaJC56XO0tHp6PVcmLJQMvnulYyNOhKFKqsiNUKK50tZPs+SPFZY9p0jJa50uRTlUol5FcciqguMRUPrvKhCUqk4O3sp1KVKS1dCZPM19M/o/fNllX4cj45GUZaW8Bk2XOZ5rCOJhU9+M9j665GJliqR+Fe5yhX0J5v+pqISPsyLLFdFqW3fE1UWOjJBQVkUcgXldKkSsy2BoBz0Q6cLKNfm5zkwJpWJwg4ik9x3N0ejSxlXCglkqfp5Pjp9X5PS0MWFKqBQ/Sitkik+uRikXyUKiAqi2K+oSnW5uaGPfqa/ZPxYIE38CEklSpqqUAkO1TwrAtK2utEN1CBnM20/fhswaU4CBDL3e6ULNYMeeq75zkNvSl3G2Ko39OZ1Pey99j1tVSt0GsXVTM3gD/H07WtRS/YYu7y2az9q47pF4y+kMVmkZQmEajkOySGxRqQniTJGLlRfbZBlrpyqPxa++NEGYSqyWA2NEAi7KVE2jYF2y7QkrsxlFfefhq15U0Evp0r82pT5dSTKTOgZjKOer6CfGjtvhzObqfgL0+lzrWZDHaWXtaR1vz/qi7TrEkp0790apti365miE8vahiHNk7DFLzT+SnQqS5RfoOYb7vuBKj8WT+MruTLAMyLz6wjZYCluBIOqiPuceip+efejr4oICah/JktJL0BuCTSzIpHESnxzSuVOrrG6m79PpVH/9fXRNb65GYufMpCtEkCSvo+yKp6JpbxUIlbpAmQ3hk6WZrxR2qpCb4dbWI19LfIMQAn/ACHCZsNYZ1UR7aj+WHOg80CflBW6t2agcXheplE2wyenI8uQXlvtY6GwICIC8ZoYDRLuJW9TOzvTINGQOO11qDuGwFDCNEY92IUD1EcvVwErqL7ntsOvQRwH4M92K5TZcIK4oxKV6lMbVOoH91IhpGI1iOhtljYo5WndiaFUTcBPJHg7O8ugSsdNnFzDnmK7HTtXP8Ecjuyx1SVge5BIXm8LEYxmjGaMZhvQrIsWDGr3BzWsuB4XeFHECrRsFtIJppHeKoitNcY3xje21vYP2NhaYzRjNGNr7aBAbRtr7ebitjLYY1Nm+aRQ0jnFux1XaZVgw1U6m2TzCTrl6Sn9anyWWVXS6+yjSs2o53l+R/TF1XLVeAEF1pyDpC5UqgqJWjsW1vsK4lyoU3EsPpxBGfvxOGjBh1PyO1syc51qs2zRoUrQe1k0DtbxMDnXhY0Up1Xpf8ILKXIJgyzjxn97fCwqU8k4Xh3B6zLLYhHBr7H1kBtViESu6BG1jgrIdEZjqtNKefm8VIWhOEXjZMaygc+025MW71sdKUD14S1JP76FXswWqiS3NxKdYHPuEdjAzyd61pcbTz+luXVgZdCTr2kBcjkxShbREh5fZfG8kMlExvlSYr3LrMyA/4QFMzXBsEeRFwr+G+rRs0rHwM1zqDs0oXyrLlBVi0WFEYrhFUvge+52fVjHt6JxS5CmeTKFRiuZtgvF8W4rrs9d98XNXrECmkP/BsM4+W0QuXZCZS/HodSGkhIiNRR4LmlYOhjnw0W16puxeCmjZfNAYBxFeArCsnO8W1Mp/NPEXNr4HbSzCY6+cSANtcTZAhR4eg+N9xQm05UFnKFP1olHd7x703ZEXxUr4axwi1rGGXO6EFC9wPrlVMeayO7EvH7rWyMLD37MFm+JQ73R7dZMEc5mRrAtHQgo+nhiv+2DFHYU3H64MZSBwSB5r2t1dH5eZAkIaISme0NZYIXKOEFFO+SEovHHNhrPkXSOpP8RkXSSseGYsk0GQQiUYKk5cy9VV6JBxdvFiAhM1i02vEliSrDxRnVMX1kyC2jkgIV7HqwybNOJIbFdMrbLOus70H0DjIX1sQQ20cdoNcJkEYH9NgWyKS3Jk6xQDiWOaF0XWJlDcfS/aI1Fnb7P4mpJcmalH3Cz/CMWWy9oZDYss6gnpH8W6lDUWhL/17uGusy+xSe0u7+iOxJth8V0Ve46/97qVVgzJGnHofCHD0q9MGorQevpHTXB03DqUJPccSXCizBehPEijBZhvXn4JdgTinyeOj3203WwBGvcouTV1miZ57IwZIC7Qqg59GwCJkBZmY1GgsJW63RSmwdHo7nUMT02VRQpBTZT32Z4XhUFQhrWhY2ytXn0CJpzl0kvwKtWXwbYuHYyrLm2LfjeKc6yH67h/yCzZX8iW2y+9AenZRawKcOmDJsybMqwKfNnmzJhdNU5bCxx9qM/DD/6nxTarttVNDFuHzeGny7IfST0vN70sxuzA2el9yhu9PG/vNzZ6rROy42Ugy12A2oW4XPUeum0CsajH05wYWzUMWjh10NbC88pd4fiDuIKrDWdXspYz8gNKmNMRiJE+kw7/HrY6GeVbg9/OD9/67EN7YZRAxKb1gcviwKRGgtjgoBZZkUpTJUkqKfecUxlrpY6WmKsEDtdaOliADAP2PdUD+lHCci8bc22NFYdZyn0TSyrRKbHMEgzOY3Rwqu/aTWHbPEU94VFW/QycPrbvDLbGefIruzchJs7Cw0CZ13aOrXJZViznGK+AlZuchXpuY5EFkU0uUSq065dpNuNsB+tmnmBvBOb1jCvaIXIXLhiLN5l1WIZr2B44ji7MuLdq+fib//95G8EX+qTRJkP5QtEPbAt5Uy8syLetGdLyXea89W3w5oTxRqnuCWmleEfMPvBx+nKE0aqIAgLDTovZJJVboOeThQrFSvVA1Oqr75t2YHiDUq60ywTqtb/ZZVNRFWfrB1rU0TBarA5qWPxFpYJ0KASpEcuJChJDC8LP1N9s26mMqq4hMHH1MsKF5KYIZaKKlWfgN9okyjHBlYtVq0HpFrfhPMVpY+ioXtmpb2p36nX077Ie2bMxuLnCiQBT2/AWWmqYOJJsxIYBu+cetUeql6G+i2EWalYqR6WUj0Zqm0LFSItQVGTC4O9eSbLaPk6nSvLlkewVsOl/lmWpY/7y9Tv3fp4IL/LJxhZc9AtfA2Mk0+tcaMALaK04rH1w8DiDgABm3Jh2QzT7bNstiJH7mc6eQWrG3QID/r67YEVGsQ6PG0CZa3CrOkmLzpxZ14UaqGNna8pD+nHWCZSnJUy+ki52fIShoUU41JLonhCBAxM9TPaakKD7Fl2mzto2+ycTcd7YCZ5y5tfPxg+1ANfrznPQ7q09CY/dKeTPOxnR4Fyu+wwG41C/mXkwIQFi0afHB/0wZuteLMVbx19WLus7pTSsDcBdkYzRrMHjmZ7kh90UKDGB30wvjG+7Qe+MbCxtcZoxmh2GGjG1tq+HvRBND8NOPdae1C1EXOtQJG1sZ567cI+8UpMVy0PKWheLo0R755+LxzxcUgBz7sgJ2uSXdoto36fp4eaqqwKNbh3kV2DDM8Mz2xs7jUus7HJaMZoxsbmQYEauwYZ3xjf9gPfGNjYWmM0YzQ7DDRja20/XYO7tL1z/GSoIEYVb+pNk92mu2zBJnWWUg9pF5JLR20Ovt3mEKdNaZJmZTwPvZgEj4ZTJW2BtcmS9vVd0yU9edy6muQgaXTlYQJN03msOsTp/GMseQntmB0hgRU9AqKRBLK40dWdLJzOYEBXqOahbxb1nxywMISZgOF1uZjdeh5h8q9LUj6iTGAL8g4/s1gAA3VKwmEes6uWp0ueLtn4f3jzJBv/jGaMZmz8HxSosauW8Y3xbT/wjYGNrTVGM0azw0AzttYO2VV7Rk7A9c7atpPQ+SFROvEIiKKi0wUNiqrL+vTa604HuL/3Fh2PbbnxD4Y9t/h6rd+WvJjhzSJrT5ntVvBLqn+vWucK+KMSmuMkqZ76ZAhNhxFu9A3X5b2E1G5x11F2svJExxMdm+0Pa4Zjs53RjNGMzfaDAjV2sjK+Mb7tB74xsLG1xmjGaHYYaMbW2uE4WesLfwbcrXg3jb8+Y3OGbKFyUFDlNNaqK4yrvVSD/I46vcyiz5YpK43ReLBvW4baT4e9rnWZNa7X2iX8hWn7X3dIm+19z75Rnp94fmJr+6FNTGxtM5oxmrG1fVCgxr5RxjfGt/3ANwY2ttYYzRjNDgPN2FrbT9/on3Sp8rvmLuXaC+ivUq5vV04XY2GPLqX98qfiWHw4gzL243HQgg+nweVPuJ3dLFt0qBL0XhZNbuZ4mJzrwkaK06oM7qSWeDF6qWWTiCqOj0VlKjxX9cg7dtGL7Hb108VMcuVzTF3GKd6phGOq00p5+ayPTbBXWPmygc+025MW71sdKUD18eo1lH58C72YLZS95Kq+hT1QvF1vPV+beevppzS3DqwMevI1LUAuJ0bJIlrC46ssnhcymcg4X+JdU/kyKzPgP2HBTE3oFNu8wBvLQj16VukYuIke+tCE8q26QFUtFlUCsjG8Ygl8z92ur7u9fQsa/ubB4FF92Xv9ZAqNVjJtF4rj3VZcn7vui5u9YgU0h/4NhnHy2yBy7YTKTbinkdpQUjrBn+eShqWDcT9q4zTNqb4Zi5cyWgbJ5RhHEZ6CsOwc32sXQBu/g3Y2N71vDlEhMQmY54DTx2ge1QGhx0FG/LojXrYwsV3bAxObLrvbCDN8hx4vQHgBwgsQdqfsycqD3SmMZoxm7E45KFDj4BfjG+PbfuAbAxtba4xmjGaHgWZsre1r8CuyB570cYHv0GN4Znj+z4BnxmU2NhnNGM0OA83Y2GTXIOMb49uh4hsDG1trjGaMZoeBZmyt7adr8B4pmaGCfNY79PyNd/fOl+TL9PgyPZ43ed58wPMmT5i8CmA0YzQ7DDTjVQD7bBnfGN8OFd8Y2NhaYzRjNDsMNGNr7ZB9tn/4ZXr28rt7O3H5Tj32tfJ8x/MdW+8PaqJj653RjNGMrfeDAjX2tTK+Mb7tB74xsLG1xmjGaHYYaMbW2uH4Wv+6O/WC6+/u7Xfl+/V4ruK5iucqtrzZ8mY0YzQ7VDRjy5v9pIxvjG+Him8MbGytMZoxmh0GmrG1tp9+Ur5fj+/XC4ad79drfM/drvP9en89K/h+vb29Xy+cjd40x6wM86g+wsRZl1HvKBk81AQaNAGclVbW25Te4nNUf4OmHvYAOEO0/KdoAQFYL1abLU1fcjiENUjwc/k6FoVSs1XLpgkeDQfqbIGhKF2vqXewTgIh+p4qOnNEzzzNbstssX7ldhFgVAyMseHDpV4sFXQDeDSVUx3rcuVmJiioUFlMqfLx51j+glbkbVOxfrJuAwq835KpaMElORoDVRE2w0FZt5LnVjwNTsizLElB+lHyPeGx+AH5UohLGVekFBEqtAqK2xkVOzBYmevjk/G33/TqrpIqBsW6VG2uL0GllxkszXA/TFpFsapM0KIXliZpF5K9nyCdZ/nbjWIEhY5z8ci15fFaeaLzjkyz38YkgKYoVUY1po6x5lM03Pt/fCc+m5R97EnZxw1S9nF7KSNqp73Zs0v/JxIF2/ncM8DiooFq7KanZnB9c4CbMM/C31/ebRVDrWsP87+2GuaP6wYXhBJ4jHBRF3Ar+Y8wL0FLY/1RxQ40zHbXuvY9Ow5h770cZfAeBu8eVDGIM4j/YSDeQ0kG8wcK5uEaolf1DUL4J+9iGmajV6EhLzF8rJMqEWnAXstaWs3IFJnYuHjqlYlfR1lPjgW8KKtsTtsqqwp3KqTIY1C+polIMs0AGT9FgLCtlZXfem0d4iTDhcqhFqxnkisYtHJ1K0h8uUZopqq8UtDA46/GT8i9BP8fi7eZ0aQ+DpuoBv3/SqTqynMB3cczQU46RadN2nMr81zJwjODGm0yMZfFkQe47qqRxlQvs2xGgy3jj0JO0Qll68p1ZMbe77iF56lxKHX2daNmiK8BLn6vVBpBp6DdhfWi4XVFVzq2HsbQiTevCts1K3DC7uNGjyIOEHaOPsNRc2tii01Ixle043q9BoCLvmi/tWtb5yy7hx9uX31trSEbGNK8XsQP9XGdl+l2XhDNyTDWb9GkZux63qhA/cqiUjfB7wGud313gz6truOr7y+zktBfqZBnbnsxDLjScWaREL6oz4m1ILz95FFLXodsSLGru8HlZS2nWXBarZCXUsdyGqvh8ykm0TLT0eB07qVcVmU2Clp7NEphkPqxiaA1/+uhD1tT2bMg/PcUSQOSCUBJ1DatyKNH7RljibEYj6EvoDF43O/s9FT8YqwTfwKoMtf2KA4lZ01v7AwysUHW2zpFzs6j0bxK6SCPSSnxbb4ql/AjBmbe2r/X7giQIJdbLDOEZR/fDZ2Xz1bCCfhR63hiwkhZovFMp4cgo2RNACdTYNsUbTk5kznOpa0hxOgFCR71+RV9NsZ+udBF0wQ6NdkdaYxBF/Se/vPs55+853SQUsiaYYqA4nFmr8kDgv/jP/iu/hLl+R8wdS2Gawi4PVwBzHY5THjQ3rdUVJgVTCafMEQkRewk3tfmub2LzHhvbCAwa8f8e1DvBKZwk6tIz7XddEAm5KLCA6rFIysTsAaaPYYm2CA2zY3QFexOeZWJzMblKCbn90EEQzcx0VIl0rFj4arU9YDZ12CrZnb9iEdjF2AWVnmeFaW1mW1fxgM1OHpD1J/99Mo/8PcYWjmEv5IMxSdWn/QUz8fGCB7O7Oh67zdBl5/DqxuqqeMJPHUN7Ktms5IIPhhaT3jlsqa6X7C0a2mWLeFzkJbw5w7T7F8wow2Z7YH8+IwOG4mAhS9mlqB6AHcE4u5rVGinc2cv/nVkmQf/ZvMS3RTiQ76aSeBr9KGeT+6wVAoZGkx7/4THZ/S0Lb9DngTQ6KpoVDHo5LFTy7qf3n61kZE/XTJdke2l0tNse8XsM2D2NJ0/QCkMcKaDpLfL5J3kC1nUcuRRvTsLlWvv7QJ1J//pBkRhrGSs3AesrIXpz5dRRs2DRs3QUdmpgVxbhZLJGg/CuqXZXJwXFZip3gOJaZtnZy+FusSMP0vSpzE2mWDhMvQVsobkPs4W6CjfynO03ie9zpn6A6wYE5l6v627hgbjPzn5GbP0MTbHOtYEtCXw2mtFCrAFltzGKSdLeLvOIHnbGuuq9W2yy9Zp4/ALfAkkBLeuovpy4n0Em7NLOv6Kzf659Z6LVhaSpdjl1A4+jHOH+FIsYdxbq3RxtdTAQkxZxDXTFNfhcZxd4c1Fr0tcqap0ZoTT5NfN0aLiFZVDb2sk82ZUwv3f4yGQuWhtV7cc2MlRU3fsqO7vEXYYeTcWz0Gd6DolKzsrn6OJa/3GieFBit7d7uRh3w37bhy3CfPp2F2fTjyZqqW81Flxq4MyRyUiRc9jGd3umrQARMjjRj27VEUBcxQNjxts0bqnzCYKh+cFP3PNGtu6gQVPc6vJSMR5RWbdM4SdCOxah+sWVPLO/mXWk2mSndc0wgZrOg3DoJ0GiagcF9yehy+ur2t5CHzKNzdfkDWWztAKUUHAqB7P0APtrJIp5jwaQHOcMVoYZYdv40TWwmw70+xk3pDm2exUnzbZSvQEpW2SPd9hlMqU1kJpqLh4RQKomcHPEUyKNE/Jcgm/Ti6/PNHpHLiRRuoEkzuPm7Anppiq4pJiNO/dFqLRsizz05MTMAKOl0BKzY5jBJljgyAzho8Jz42KqkJjJBM+fOEZ9/4C3yFouMY+y2arl/YWO1SXur+BytQdR0r2BJQR5tQHmbDdDTmfgkc43biv6osEe98FhdwJznV6e2+vz+BX4RkrA61qbbTo7KcIk/HfXw9VHIT4uvG7IKO9Kd/Njw7icheD6bVhDqxDLJfL1Am8PxkOVn/pQ7rvfVUXTWDz/fVtHWiDXt2FVnTQ7mQJLSlvJAxNwn6uHXCUu761V3TtZf9NYMOj5nQMamsyP+lbjzs3b+2kEYILaGZG1J3Ch9P5WbioAoKoppaLl1/2czWcuUdI2VqNOeMspEy2gMNi82sqOv9In6cDraQ9d4ayDYCKxpPlC8t4d+48DOHcwT8QhVn2V4ytx1lB5kWfONkfMSx8CgR526o3iAp2qkXwgvUH8sTJ0/f1msmnpnv0qk97X2hYRzV551Wdt1CvJbxt2osRN7r8metxoohYdQITpKZQqdujaaH5vR3HGpwRWTrwDGqBEIxlr68xieOXIr65wccArgViL/x5KQuNwVRCX/hAyRmAJ6rlR7Wyy2Ds4vG5XUpTmoiVwtjJyAmJLKKd/eJpFKm8vLXsRTDhvP357BzX6JlNu0QWwNNCXqGSwn9PRxh9sUEmLEDPrkexTBcVbW21KnOD//wbNHPxEQ== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate a chat completion for the given messages using the specified model. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams"},"tools":{"type":"array","items":{"type":"object","properties":{"tool_name":{"oneOf":[{"type":"string","enum":["brave_search","wolfram_alpha","photogen","code_interpreter"],"title":"BuiltinTool"},{"type":"string"}]},"description":{"type":"string"},"parameters":{"type":"object","additionalProperties":{"type":"object","properties":{"param_type":{"type":"string"},"description":{"type":"string"},"required":{"type":"boolean","default":true},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}},"additionalProperties":false,"required":["param_type"],"title":"ToolParamDefinition"}}},"additionalProperties":false,"required":["tool_name"],"title":"ToolDefinition"},"description":"(Optional) List of tool definitions available to the model."},"tool_choice":{"type":"string","enum":["auto","required","none"],"description":"(Optional) Whether tool use is required or automatic. Defaults to ToolChoice.auto. .. deprecated:: Use tool_config instead."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls. .. deprecated:: Use tool_config instead."},"response_format":{"description":"(Optional) Grammar specification for guided (structured) decoding. There are two options: - `ResponseFormat.json_schema`: The grammar is a JSON schema. Most providers support this format. - `ResponseFormat.grammar`: The grammar is a BNF grammar. This format is more flexible, but not all providers support it.","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"},"stream":{"type":"boolean","description":"(Optional) If True, generate an SSE event stream of the response. Defaults to False."},"logprobs":{"type":"object","properties":{"top_k":{"type":"integer","default":0,"description":"How many tokens (for each position) to return log probabilities for."}},"additionalProperties":false,"description":"(Optional) If specified, log probabilities for each token position will be returned."},"tool_config":{"description":"(Optional) Configuration for tool use.","type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig"}},"additionalProperties":false,"required":["model_id","messages"],"title":"ChatCompletionRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-a-completion-for-the-given-content-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-a-completion-for-the-given-content-using-the-specified-model.api.mdx new file mode 100644 index 0000000..3e826fa --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-a-completion-for-the-given-content-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-a-completion-for-the-given-content-using-the-specified-model +title: "Generate a completion for the given content using the specified model." +description: "Generate a completion for the given content using the specified model." +sidebar_label: "Generate a completion for the given content using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztXGtv2zgW/SuEv2wLJE5atDvbLGaB9J2ZPoI4/bBoggwt0TYnlKghqaTewP99772kLNqSH3nMYFqoaJtYou6L5x6SV6RvekbYQudW2N7BTe/p/j7+SIVNjCyc1HnvoHc0YtYZwbOf33JlxQ4zwpUmt4yzVzorlMB2J0EMu5ZuwtxEsFGpFEvmDfqslnNqylhMzgaDN0xcidyFFkyPWmQP6N6rSZlf9ns7vUTnDh5Bg3lRKJlwbLz3u0Wrb3o2mYiM429uWgjwQw9/F4mDBwujC2Gc9D5nwhmZ2KghN4ZPoZ10IrPbCojagQ8yH0O7xTieQlByngl0DgMUnpvt9K64KgUK0Ln4POodfJ2LkuDiWBhoNL+Ul9kQr5y3yi8zAUIZSWwqKnPpNhv66DP9xtVjhjLxIRSVCW5LIzLsppE2kWivrjcDDTxNpX/4OArSiIADWPujlEak4GEVtcp58MZJp9Cqj3TnaN7tDfsOF9QymSeqTEUKv7DD4yM2R3QfXV7p2gdpg1vU/4xbqxPJHQiaYzgWh8IiyG3u67HIhSF5dRYwJ745lGSdLi4AzgGsy9IEdCRGSeTphR5dYKLQVfqUCWv5GOOiS0e39aXIba+BiBOSz64n08oYNAFVF9AJYIXSY8Dy8B7gryRcDKfejLZH2kHRQPSy+a9lQk+ZKcsgwyEyzHsKP7B7pGGgnqF+PpRKkthbgbBpfYTDU7zwQY+PKULLxn1Y1kw5Ufe5t7QFghUCm7a3SridQxVAFwEWOdVk1YZncyofGZ0ByUfoRVXCuj4ZhVDeI9be9azdsW7Hun8G66ZCOb45gJ/ENQv4j7LIgqGCKQ5qE5o5QHSlZQnMOobQVzn8MyzTRsQZ+5ew85wIzDqa3mFyVKWg6Dj7B+RsD++1HB3NfFvyknCNScXD9HlxylGlErI2GvasbZKPlBPYnV1zC1kNaS5TSg6uwEkQ+nBzbuu4K2M4VYy77Nz709Nj5luDSyklQAjTJjp4YwzSJTbewchMtIHlRZlliM3A0YLaXE9kMmHSO20kD0wLDOHvkx7ChONSbavZt0bVSufgG5uUGc93oX9SPlSCRc8smIOaZA4+A2/datQ4ZF9OPgRnAr2VljCJSy0jYaj2RCdz7E/PMnwILEXKbSESOQJ+10lSGiOQNhftug2oQw9XvTUPXgRzCtOK4JmFOUgYFPrsRJfjiZpC9yilry07efuK/fSv/Z8oZcU3jpiP8QVQr/W95Ck78RCv7dkS+SFznr5oz5xESRx1JvC4xV+c1vBwPq0Eo1QAwljCbInxTJc5DYJOZqJLqi6pvrOkevoiHnA1+4hID5ll49T6ry4hJyBC4lsiBE4WUSmOlEzJTLo+O1Yw94GpF6CHjzkkiYKbphqpnq8aqawwV9D5EE5MJQHhQWCVufgG8cZxWIQwdKnVpdZ3lFrP4/HqCJGNk86BR3utP6TXYRPyVTDSPvtcAhKwloqj0lDAwJNrBwGDeyG9UjHipXLNFFsnuEuqLqm+r6Tab9O2RQpRliDU+NiiNy+5SyZH+Uj4sDyCZRoubwda54+bq/t3YU24WECrKjh+JlhVK0pL6+Q6/mBLBkBX+Jqj4AZWdEAFaMS5DzAMtC91OkUvH+otCKq7kOl2BTWZgkq008yrXfg8ogzQ1qcClr+SlTCtHuIMeyytH6mpzvNB8YyzgePJJeQYDN9X0CGUEleSk8Q9EmBhkE8LDdTQXyrAN6p2wV4s2jXurfDa3216DGpgkQCMlPlKzpwpoyuLVQ/8BHDgLlBFFZeqi7E002eH6ppPLTvzYs56lOYkcWMHlUa1EDVle1BFgnDZknLH6QbMp+hNHCU6NsM7B9Ru7wbVzf49hOnXP5/t3OCdWZ990oBYN4HW+DzMrlTqZ3BK5GPoNpq2WYTlJmvldkhCNf42o172BSM/TxRWlya5ZYUTNUf8AvKJtXl7BTF0zxuYSKZ45aDnA9Kw1V+mGScVPSnWFGhc8wXkbbJz6b0qyeD4HnVJegiIH37Ig9tEgHyscBWFgvS98g4fARxbKknerRizcQ38jklEr7ziHJpfuFcKoRSfQSRvM9qgVSXrTgElNXGtET6vD6eLVIZo4jgRe4p2h3BOPwHPY3i8ulA+xQZbUkTHZh2bdWy2gc2W2aIjtfuT2my2tGpVArIsjeVEw0jjbVU3W+v4reO3brb29yO2brbWsVnHZt1s7YcitW1ma7PzdW1as6lSAhk0XlX79JursEIL4bmgyqZtMktUzj6eFz9RLqowWvlKaZCCgAVV4+lmYqharnin2CbwoaamYyNEOl2AYHQpNuVjKJiehQZnPfQ8FFyn4Lq0TVPvAKaog9+RokEQOqhkNira1Kyp3HO2FQoCY6lvJnI8we0U9U6pqd8ZxaCh4MkEnhVF/yFmK7q4KBYze36lPap0f8ugYsJlBUK5NLEZYfPbspJXHp6WGZ6nOsuFpX0fleA+e49xMX4DJd5KDL39rpuTSnKgVVnwcb//4nlDd5mVijsJw+VC1Ccwkk00jKT45iEvEyVKG1n02suk7EKx9wPSqS6ON8IIGu0W7FGw5fFKPNF7HFu/+7EZVwpRZYXfPOO3FdLbhKTd+//8zB4MZZcNlDW34i2i7HJ7lJG0zS9CPxEUvPNFtK8S7QQ1hjq57tzKHIgmbrNw7MndBh2ybrGbf92qmy9XdS6AErf/uigeYeJ1yTINlip5KdS03pB5p4l4YNh7zx468m4n7wZVdSTekfifRuINluzI/Dsl83gB0lA9Qwr/Vh2aaA9jlUJtyxB4WGZlxvIovD60aGvYElNv2A9livicKkrxhJdUG3WnujQIyqwAXlSQfLWJKDLXLmxxrPdC/MP6FdE3F+oXhGEjCtCCei4KAZ3mpmtJ4skK0AyFu8YdXbtP+/u0fQJ+9tmxtpLSJ3ATaZD/EywX11UUcLWfMjptIhzyGPw3ZTBmCm6qYJDRVrMRNzsVwVUbVCrfqE/lRGvaWuS4ugy7iLyuQib1SZrIxVUnVeZ7N5b8/YKZwZ4BXfxR4q4bi3bjWZ1wTOhaKkUnZeYnZ8DOUWm8ax5wzO/wxlM62EHoHD2GvYZdxIPbJKZS1L/tTqcA3/MmtI/9Mpd6329ouvBVsbUr33fwUAZ9Uu3I4vP9Q+NSYqHmEegsExzU0scsFb6I9BBr1up0E24hugj7h3ZgQkP2NDc61TwZPdDGlqEUGIioouNFLTUpx9chdPHHliFg1WmlRizyUql4RTDUGug0bzvqt9NE5xKOmyuL2XkDwoi+XwafPzFvfVVe9DvbYFjHMicEA6MDMeuzIzwpcDx1E+juwetfd3zw4K8eOZyEsd+Kacohrslv1SaxuwwEcUAj0P4Clwd0tTp79NZjtWWeNJLj0tS4jJzcDRid+1mfa7vPeuzOyAxNtkdlJXNxzu+vQbCH+eg7ROHLT2/ZuJVW1mPyTvjCEC0sU0jvrUEV7F0PqDutDjcwSseVHVf+HbhyDqa/HqMda/7QrBkvw5Y00MS9+kqHpWivOXxwNGL+q33qF0Dt3+0TR26xfkFfMdRfcc59FfDXrLhXLRXf62t/QjOsix5h2Ki6VdAqSuePw5kJWDa0H/vegkvWRWq+1X9nxalyssYvRCub/MJlWC9nbsln813+9U76CG3xuXN/TJYwUguABYfA9TmsHDV87IFZ1BvcTeDT3tWTPVkdj9irV9TQwp/Xs5TMtJmgN3GuONjbgx7YnYAUke4qPBGwa/FEQB8eplHdiqQ0EhfJ8ODrqjO/nuM9pLOT+izEm/rgR32WofGyPb7U8h4yfkMYwBQqvUtlif32pfyTasH7tVJz3r7sC8IXKXlx3J5FSYiBX8oIj/n9GR0UGmm6FvoxPlsxiBkLdGA3eDRePWl5jetHz/gkkCeVkLOxZKqngD1XQDH2LGdLf3hV4qvOb1gqVPivgEi08UHBNSvDfDUjjqt7EsqNOMNludKGTi41hQ+xhKiAVQzuGfBWfaTjIkQdiEtIboxJ7qd/D3YaZyFg0embh9MQoEHfrlMoLnP6Chm/A8dn2lffd/NcI3jPsw1SAzMKm93cYLnni1GzGV6GXDFTf4iIjrgNES1fEaETwVMgTMyySzENVXD8Yp9TPyyFr8ppnjDCUdQ/cZgkonBr255H1HH8eXCK4104xYTew1XDr+Ei/n/Qw0oGxZnwTtdueorn45L2LPW8TPzzf6CyyPA= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate a completion for the given content using the specified model. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams"},"response_format":{"description":"(Optional) Grammar specification for guided (structured) decoding.","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"},"stream":{"type":"boolean","description":"(Optional) If True, generate an SSE event stream of the response. Defaults to False."},"logprobs":{"type":"object","properties":{"top_k":{"type":"integer","default":0,"description":"How many tokens (for each position) to return log probabilities for."}},"additionalProperties":false,"description":"(Optional) If specified, log probabilities for each token position will be returned."}},"additionalProperties":false,"required":["model_id","content"],"title":"CompletionRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-an-open-ai-compatible-chat-completion-for-the-given-messages-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-an-open-ai-compatible-chat-completion-for-the-given-messages-using-the-specified-model.api.mdx new file mode 100644 index 0000000..c3034a7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-an-open-ai-compatible-chat-completion-for-the-given-messages-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-an-open-ai-compatible-chat-completion-for-the-given-messages-using-the-specified-model +title: "Generate an OpenAI-compatible chat completion for the given messages using the specified model." +description: "Generate an OpenAI-compatible chat completion for the given messages using the specified model." +sidebar_label: "Generate an OpenAI-compatible chat completion for the given messages using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztHWtv4zbyrwj+clsg67jFFW33W5rt3uWw2wSbLA6HZGHQEm2zq9eJUhJfkP9+M3xI1MO25CStk0zRNolEDofDeXNE3o0yLtMkllyO3t2NfphM8EfApZ+JNBdJPHo3Ooq905THRyfHS5YfJ1EacnwzHh2M/CTOeZxjH5amofAZvjn8Q2LHu5H0lzxi+FsS89P56N3l3ShfpRyAJrM/uJ8DiDRLUp7lQiMgAvy/aSPzTMQLaFPH52LJvZP3XjL3cvjNB6Q8v8RqdA9YLRPha3gGEssytgJAIueR+3wNFhGXki14mxQ4tHnpzbMkUhhEScBDgNF3jlkS8q5ZAjElkHJUSJ6pSc9ZEdYeuKh8KmTuzbh3pV5fjbw88UQAiyHmK0BLSI/Bvx6+tCgr2lQr1sLXYHJ/sJ5ufeeo366fY85v89ocywfdc8TXa+aIrzw7LZihgtSLidyelp1KUgEkFgQCm7PwzJncnIWSH4Dc/LcQGQd+vdRDmZG/wk+R4wqPuqTmWA93xrL8AprDTxa1kXMRS6GpN08yI4RvkddBzGZhi/ct8nLsruGOKyQigDQtsrC2TPWn3WtVtlmzYOq9u2IV0JbAnai2Xz5/9GTKfTE3GsZjceDBfEDKJaDtBTxnIpS4BhsnbcbYzBo4muEGjSpOIvbDIoAHcZ1LsC8OvR3qm9NU89J33kd+zUMcQvdVi6tHqqY09o5ZrOkZJjdXowP4ZSkWS/wNml+NWJEnV6NhfIrTb/GnIjHMeSeWr5auP9+rAbsZ/8Rljb+K8+ci5DWmVw+AOOrnVuOBraYBy1l7DANj2mXlzLuYRR3IbV2aOuk/AKAPCukdllTNsrWaGhw8DgQuWCRilicZYmpmv/pdYW6hROAPIOY4lbpCJouxpxajrofJepD1eI3Wo6eWJyOysxFBeD0XW6/zvbI7LWXcrYcPvJul8JeeDwJg5U6pbRT6BBpiOBQIhrTqptMG2cNhsZMd042uFMMJX6QszsfDiKYiwio2a9HtCwzzSY/SzflH7ahUoSZQ1/UQAESGy7yP97QtepUrCcFijfWdR906XzdYG8Pq148YxZInsteeSC9xt1yDVi9KgcBj72TuRbAiAkA2eAaInSn7eA1LEBwggJV6BEDBJwCFmfNgrMS7iAOehSv0Cz6GLGLeec78b9AQ9EjEoFcoEw+oa3RJc5w3OFt+y3BeB2rq8F/E8hwB5kkSgtmei1itjfxu/ChaqI7DE+qhczXQFk3UxEYRHWcvQHyywlczRy9EDQPMAoJRptEeQwMxcIFkDrOvCWn9abeklm3WiatF82/Ss3lTUkekjtawxYMFu+THNbIN9AGNMvVZGD4g2y1A4906zQRMbNGRcXYQPcEuFkul1RAHG9+EgLSKzHok8x2gX2IBXojlSsGV6qyPoKa8zfMuYqVj6t63+7BbLmyTtW6IbVBiUgJtBZ7OrD64vXqHmoNZ53eHbUpEYR4WV5YtigimJAdBPbK9EFQK7GiVdTmEosy/zk9/93aLMuoifQELfQwYW6LtoDu2KA07QltRlFxczxX0UxflLtpYR/Uu5I8gDmpp7Ahy7P3GIERxBEdWfno3vp5mlF0MeosiR1at9IwpUNUx8FnQg3FU3ZtSPX1Xzv+viDiQjHUDaB+sMYDweq2QqzVxlXipYTsD7lb2Z7sO04NVBJPYhLyI1+lFlGxggZfxu2LiXbz3GsNucOZRs/RUABlPAVFcZqMC4C/gElwiIzEivk6MwvwLNECAKVHsUOPg+tNuNi7brFUIZQvKP7xuUW1MtpMvHuzxt6A+YTT/3o41NLVYITlc1nvsIBq8nR1EVQCzVW6oroZ0A9XV0M4o7YxSXU3r3X5viVJdzQu2GFRXQ9aDrAfV1VBdzWuuq7FVMA8OY6nAhlwSKrChApvdC2ycWpgHayMqtiHVRMU2VGxDxTZUbEPFNvtfbKOLYx5s9qnmhpyJ5+pMUM2N0Zu2PObB2oDqb0hsqf7mWdTftJPEtdFVojsWcjnNODOn+GxnRt228lfAPU3SlAfegsc8Y7lJofeNfRCkcIMefaAPggiTBQjrTG6MArA/NMTMyIzNRChwiRyn4BuPZXPPaFuM4GicHbUKDtu5szBb5b0OK7Lkuq/o4LyNi2jGM+09pVOXTvuN7yBdpXCqoHXY1/RjsjjDd51KM0kbjLFSbME0U2gJ6ylQ2ocwErXzDBqr1TEhaL9pSo8+nS3Kcqhc3eP854Vk4f6zIokOic4+ic7Q5A4ayY+WHoNoW1nBuvW3NrvDMVQWue2saFP9OPSzuRx7nh80MBK3wTtH+GPnPEDXUe961+QKPYKH4O2W940IQ+3KNwBgpQeMDMTKedDTsfkSi1svF0DxnEUprrnkgHsgYTAedx1p6N1ApGAHgfH0eYO9HDPtjeUID4EUEnwyiEOMW8a7D1Acwjg63DXrc1ApRIuuxXZLZNFC/rON0YcwUu/vSfbjlEmgi65DaQ+sXlUOv78s4m87uKj94p9WYIVoddrtfvBMvza8NVmEzcCgUwsSbcTQRgxtxOz3RkxDjqv9mIY07+LkvLeKgHIVlKuggIsCLspV7Du+JDovW3T+vFyF9hsGZCogdtqQroC3w5MWICWcRbgz+8Tpi7EN/dYmMZwWD0llaDCU0HikhMZx56Kop4bNOngIZzL0c/p7pZj+3nVvh/Z2VUtFLBFfs1AEWBkcsRCrprlbrPBu2x0emw0JrH5edNmFJhH+eXFx5unWquh7VO1IbuOA37IMtRA2PkAKLpMs92QRRSxbWf+bqzaa24WedCawqlOpzNi8V+Pc9/9OTI+sW+PQYRLD3LxlEbH4LaxjwHCdnD41dHRAgdVb/rD49kh9ZFd9eDPjJctnHHrzaxSJDEMQXQWvgtNZUuS6bt3EdF7i+0WWcRi+gdcQ0TArbFerJJ4jCopMa4iX1VJ6iMPR2cnY+5wUi2WIFi0Mkxvpff5w7P308+QnpVBNnb/LX8Dq1Xi/ssD7rFm8wqcn5xvJ+eGXbsnxQ4FJsSV0xyocDGChc7yygJUJZN5CXIOSY1FSmAwaKEISKhKqZyZUP/xSjXcBnP4JOd1IlnRF6z9JATIBFOK3PucBkg2FDW1vKCKRj72zEDwysGXAPWzBQEhCeJmNjbz9uM5SSZ5dw+IDOVGUOJAHGauI+S3QGyyvIS6JFonWsxKtH117dYKcDZh755rbq/GNeB21Wd4SIxh7pwVwAjiNyirNOBieOFGbBoEVrzJM6LjEbS1gEioSquclVJOu0XqIkJISZDW2kDibk3jOFUVGrZLTf9iQsldEZrMj2h0sP1YtpK3eNqsAGJnPMEF0sOKSg0JAVL5qMoO5/TUJVo0s/kNEckBw7exBuh/aIaMBw+mvd/WTyOwfZnwhpDbWNyJf1j7oxXMB2DWsiZKKa8EUxEMFQIKdD9JEmA/rLL02ZRvpdkWqUN+bCnU6BY7O8Xml5/g8lPPpAB86Be41Wgw6BY6sB1kPOgWOToHr/THmCzwF7qHRKx3+Rp4IHf5Gh7/R7Yqkjl6ROqID3+g7I/rOiL4zogPf6HZFOumNvAg66Y1uVyRRfUmiSqe70e2KVFfz2nUD1dXQzugr3RmlLVGqqyGLQXU1ZD3IelBdzf4ZEaqr2ee6GrpdkVySfXFJqMDmVRfY0O2KpJqetWqiYhsqtqFiGyq2oWIbul2Ram7ImaCaG7pdkcT2hYst1d/8ObcrdvtTZf7FhB0wq2ueSaV+dHpfjRH7qylAZWG+ctbBHLG9ZR1MP8UvoP3UabbmMOexGwEoRdtT9EsR7161DhhxoXz48u8ZqFrOYvdRecL5Vp3TxgTPw22xd4MQtdjHnm3kUmBQ/LlPM9+0/WYYrZyjO/EwWYh8OhNs0AybSG+muhrDwzEaI6eN8+stXbbDaxxQ7oCN2O20ks+p5vJhaQF11hW7FVEReXqKOu5Rh587J06X4z35IPFw2BXMihptwHgOWRjycNqZgemxHv9ecpVMVoG1hiX+5175oodRDpfPn0yDWc9zqlPWG1MZNU9Vty8Z6GAf6gXjAKN2mHSIR6gqf6SB72MkFOw1ax8UxIsu7C46hu6dSdDnxz5GeRMeezc1Z965FG0+7yas06pBX53TL/AEO5X6aRH5oDZEi6eQkXTOSDXoyLjkDqs9UrLMzZAZxJr6d4CX9r560QaK3Ws3FgzTBxyPpPQVmQGMx4KlPm/RJt+cgdaeZ7hHJnaLhnIZodonGyaoigNagoqQzw2tdog/XBbeogSckZqzdae3Jxqh7NvQCM7zDRpBt2poBGUbha8X01yh8eSqFwc77cZ5Ax7Dqf48SyFfmjFsmhQyjWQayTSSaazrh9ZlUWQo/zxD2c7k1fEoE3qS97opqyEO2MvNWOAlpA/Mt5eb1tsSYDiWk2SoUMD7qHZWZ7p7TaU7YKdJ2kytPUMlZuZo5uISD9YixYwKWMvBeQ2nbw2kysboi+tecDpWZ4n0BXyN2b/ERKydcp17uq/y7KlL2tddNtKhCD4dzpYANnXBNL4T7rf9owq6SyBDrIG+uMC5HKBhGZio75/Ze8CU6q4AgRPLMUULiiqBP0dpou4KAxOwhL8Or78/TBQw/A3twaGTKh2heseLWqRiIPUlzWiZ5+m7w0MWr94uARYP3oZYaPtWYqEt3qiooizJ/SITmOWEju+tzb38iu/QpH6uLnv4rbrfwlzWUNG2uhnh0m4U2w+ry13aqrV2kV1mNV3KEvYhndwy06396vV7l2Up3kTXzjnNNPc4VWxu7VkTqFPr5Rg4B0lTGVOvaan5QHW8na7uDvp2unzt3ISbtDbOKgDOftLlHfav7bPUdj+QSdduW0zqGwwTtRUwWZO315DaufZJZ37cLIaKxEtfZmLdkWoq1jfQ0FsmvWkAJw3rVWeSkh51tTcpFdXE6hrnAyG820ZhbDSAW9x+7gaWMAgKrFZG1993bO/jhUCydjUOcwM8F7K6aNTU3cur2Gv8AzEtV9tq9jYTqb7XASgC68MzTXFTPZ7zbM581MwIlGX8CkvRwiRTnmAb+AxvLwyBRzPcntZYfVKXp+jtFNA9EXOF5ulvrGmE0ZXQPP3QDqceqjSQskj600atyy/1amttjlyAf+HQSrwrnQ6Mh3obO9zdzZjkX7Lw/h4fg3BnK30Xj7ovaoacdole4ZKzABkS9No3Dk1Gpljj7YUOIq9ZWCiV2byoB9WN7nHk+zzNN7b96pips9PzC2g8M5cBIR3gacZu4CH+/91IzbV0qtWzu1HI4kWB5Ro6vNP3Hf0flGDKJw== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate an OpenAI-compatible chat completion for the given messages using the specified model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-an-open-ai-compatible-completion-for-the-given-prompt-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-an-open-ai-compatible-completion-for-the-given-prompt-using-the-specified-model.api.mdx new file mode 100644 index 0000000..8056e0b --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-an-open-ai-compatible-completion-for-the-given-prompt-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-an-open-ai-compatible-completion-for-the-given-prompt-using-the-specified-model +title: "Generate an OpenAI-compatible completion for the given prompt using the specified model." +description: "Generate an OpenAI-compatible completion for the given prompt using the specified model." +sidebar_label: "Generate an OpenAI-compatible completion for the given prompt using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztWm1P3DgQ/itWPt1JsGyrq3rlG+V6d0hUIKA6nShaeRNn49axU9tZWCH+e2dsZ+NNArvbFqmVlg+8JPZ4Xp5nZjzLfaKZqZQ0zCSH98nL8Rh/ZMykmleWK5kcJkeSnFVMHp0cq7ISDJ+Okr0kVdIyaXE9rSrBU4pvDj4Z3HSfmLRgJcXf7KJiIEZNP7HUwsZKq4ppy/2RPIvWGKu5nCUPIL1QPPUrwjuqNV3Adm5ZadbLzbnkpphoRoNCvSMsu7ODL7jM2F30hoOhM6bxlVAzOGa6wfmRe77NAKs+s2HFpwu7kWeWereKR29lXU69UVZVkwHDflJ94SXNMo5go+I8UiGnwrA9APSXmmsGsLoOOrXSbsBYbgWK85C+UtWpmp3ju70O7K8KRsAzBPYS3EynXHC7ILnShBInmeRalYQ2/NhPgSDAgqlgJC2oJemSMKSh2QgM+B4LOtEaMAjWP2XSDzene8hvZ5W37HfSPw/MdCda51w40xAu3V8lM4bOGMrTLK8NFT8/FHfU2VHnZ6LOOu+s2nvsSuxp44/1Fq/V4BuDukVMV6t6qOFNxe5FtG1YvK09G4+I7zOeUvwxnVNQwrJsuE8oVcbEIKNDCui9cg2VgRfOpkl7rFM6p7UYereV86DVahur1oBG26VuT/ix58GL4JNNPfilZsY6JoPmfwy1m4i7sI7cUoTYnAqeEQBeSQXgr3Q6/6D201hq66FM2jX036urc+JXg0WZK1TBS/1Qrm59pzWyBhfvQcIwhdKWmLosqV4QlTsGMbfmtuBpQbg3WnMqrU8yMrx357i0YSkfwNfwyX41Hi2UBNtIUZdU7kP4M4oBivasqON7YbBZphtYGSWvI/Lh4jQYk4LyU0ZqwzJIGhBZ2M3mkDKUZmAmxtNFjtCpqq073FQs5TlPiUrTWmsGx3f02gb1IcJNtJbOi1Du3PSI8/QKwFGHo/OTEblQ9awQWAOEULeGXPx9TF7/OX7tsjq7owj5GF8A9fa8tzQjFx7irT4bIj8w5+WbYeakggMrSAHbDf5ilYLNctEIdimbkhmfQ9GipaphDfqWl2xHqh2pfjFSvXzTnncFSH+PSA/MMjG1/lc1cAI8xO5SxjJ0G5IN6h8RvOR2RM4F9BTQ1wB66IwCSQS81E2levVYpTJMzyH44E6kEgP3ILBqye7A31Bdg3N31NpR65ei1qu4Xp0gskFzcunR3p4f6HXUh3zjjGxEzmpAAqOlq0pTBoVHKgsOg3eBXssGd2D2+KjgHal2pPq1SDUeOm0DCjmWINTozKA1JzJnziOo/6q+/zDJXFlbdxVrLvK+EwQ+lJWFAEFQY/eDKu5yiAP/impaMsgEqMON9y/U2bcqW3QG3t/DxUduzkPTCZ7BeaikbpDhNiPCAGkjN0PxT8oaWuopdtczbnyVvuW2IKeClpRcWpp+BodB6Z5DMBwd5pw6iQdOgIECn1UK0oILrfcWKqkkO8vBHf2L/tqRXbNyg6XRdG/t2vUyHnqgQUcFBIDvZksIdfDibJ9CyCcq3yBZdqZZfgiJkWqlmvg8z5q0UJHsqVLQl/XHDpHs/woGgdIoCTe7qHljnMDcwVSmiwmwgQq76M9F1yge9jnGaFa5oUmYf43CZ0LcTqacDk56h1PUwCh53SiQA4LhjAbdj3watYHDhgd7kdiS3k28fdtHGfbysi6jaIdJYTfQ8rkQVEGyxtz4bPE2bHDqt0Yq7oqdbKyqflQG6RG6ezYO49s4tCpA9S+/mW1+u79IxTNt/3yi3J6tONHzhayFiD3RKBg9WhJorQOjB0GTh5t1vAs2Blti50EsKgRdrdnWCIv2rohU1aTaXhjEtorFwE+9VaeGUnBTLGRWQ3XNJn5avEF5aatZqI5DH4tFA3JT5zm/21pNvw0wRy220LXIsKxDvwH12fML4ZhG/6iwVZvYDMJDge8MwilvB+HN+M71Za0Iq2uGGRRIouDPpFJuxAf9VwF/HcxfHCgnCH+LkliCWQWvVsahv9bYERbWVocHB1Qu9gsQw7J9ge3KvsF2ZQS7EyS+YdAjckxzsPGv5h51fYPvsPe6aLu0d21HGrqs1utNS9M+WRb6cVOV0bjBgjruFMGV0uR3xTVl7LL/eChVj5v0Om4yZKtQk668wF6W6XJyvCTUuOFEK6sD7+vmzc0AfMctWlf+SST3jUoASNxKXoa7Cw0fmWBgPZznLwY6WbzqmZVLD42vYbFkB3zQaw7qm4+SdL7gXsncPLdpV41ra0EK18AK7fN0hk0+clHnNMXSj0KpZh8l3kWUdkTqC0dAEAHXNo0f9Xmt3rvu2NddwCjcKNEnEu4Jz3sXWXFhdPd4zjNDosKP4A4qQbl0yVa7y4on+LWPr6c4xv2Fm1G0NAd4IZVx5f39lBr2QYuHB3wMlNILf69yl/4pguoa61MBN3QEL9D7M4MlybE3dv8K9cHlonaZuXvpwnrndxylKavsk2tvoqR1fnZ5hQkgXOzQAfBU01t4iN8PE2fksry7Z/eJoHJW4+fQh4mXiV9fAUbycpA= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate an OpenAI-compatible completion for the given prompt using the specified model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-chat-completions-for-a-batch-of-messages-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-chat-completions-for-a-batch-of-messages-using-the-specified-model.api.mdx new file mode 100644 index 0000000..c9a4c8d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-chat-completions-for-a-batch-of-messages-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-chat-completions-for-a-batch-of-messages-using-the-specified-model +title: "Generate chat completions for a batch of messages using the specified model." +description: "Generate chat completions for a batch of messages using the specified model." +sidebar_label: "Generate chat completions for a batch of messages using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztXW1zGzeS/isofoldJVF2Krnd6G63ypYdx7uxrZKcurqyVTQ4A5ITz1sGM5R4Kv33624AM5gXiqSl3Ui8TqUScQZoAI3uB40HGOB6VCidZ6lWenR8Pfr+2TP8X6h0UER5GWXp6Hj0QryUZbA4WcjyJEvyWOHzM5tNXEblQpQLJWZVHIugTqDHo4NRkKWlSkuUKfM8jgKJr45+1yj4eqSDhUok/lWucgVFZdPfVVBCxrzIclWUkanWFMv3ksmikCtIFZUq0ZuzJ6osokDfVYCXTsPvdA7p2pr6CGpIZaJENiOV2Hw3B6OljCuFArJUfZiNjj/VoiLQ0FwVkKh+lFbJFJ9cDMqvEgVCBUnsF1SlUbm5ok8+0F8yfipQJmZCUYmSuipUAl0mZlnhiTbFjW6gBBmGkcl86ilpJmOtDsCa/qiiQoXQQqc113hoTRmVMdbqHb15WxtRr34vWsWKKA3iKlQh/CFenL4Vtc2Osclrm/ZrpG2zqP+F1DoLIlmCoNpqfXEorDHgSaK0lnPVdwhUmU2n6szCJT/YYExFFquhHgJf0SXaptZQb5mWpJWZrOL+U7827ypo5VSJz02azyNRZiIKoR+j2QraGWHjTXdmoYq/050211461FB6WVtaN/9B36Rtk9Cie+/WKMW8Xa+UKDGqbRTSPPFr/CrCX0mUyhLtF6S5irt2oMuPxYv4Uq406IzEfB6hGozEjWBQFfEQRv529qsrigQJKD+UpaQXYLcEkVmRSFIlvjmmdEfXWNzNf06lVv/xw8E1vrkZi/cZ2FYJkEv5g6yKQ7GQSyVilc7BdmNoZEkYu6m20Xa4hcWY1yLPAJTwDzAirDb0dVYVwY7ujyV7Pg/yyVmheWs6GrvndRpkIT45HhmF9OpqHguFCRERSNekaLBwZ3mb6tkW+ZZkQHYpOtKtQqArZSqoBbtogNro7MpTBZV3Yhr8FsxxAP5Ms3yb9QeIb3SiUl21QaV+cCcXQinGg0jeZmuDVE7WNymUivH0iQJvV2fpFWm1iYOr31Kst1Xn6j2M4ageU1wCoQuZ5PW2EMFoxmjGaLYBzbpowaB2d1DDgut+gRdFrMDLQl+ON4z0ZkEcrTG+Mb5xtPbwgI2jNUYzRjOO1vYK1LaJ1m4ubkuDLdZllk8KJS2n3m24SqsEK67ScJLNJmVVpPSUfjWcZVaV9Dr7qlI96jHPZyRfXC5WDQsosOQcLHWuUlVI9NqxMOwrmHOhjsWh+HIOaUzmsVeDL8fEOxsxsyiN9KIlhwpB9rJoCNbxsDjbhI0Sp1XpfsILKXIJnSzjhr89PBSVrmQcrw7gdZllsQjg19gw5FoVIpErekS1owQyDalPo7RSzj6XqtC0zNGQzJjW40y7LWnpvtWQAlwf3pL141toRThXJdHeKHSC1bnDwgZmn0Rh326c/JTG1oGZQc++pgXY5UQrWQQLeHyZxbNCJhMZ5wuJ5S6yMgP9ExaEaoLLHkVeKPiv70cvqygGbX6Esv0QytXqAl21mFe4QjE8Y/G4527Th318Kxm3LNI0T6ZQaSXTdqI43m3Gdd9lX9w8KFVAdehfrxsnvw8i106o7OzYt1rfUnykhgQnkrqlg3Fuuah2fT0Wr2WwaB4IXEcRToIw6hzvVlVa/mnWXNr47dWzWWp9Z0EaSomzOTjw9A4e7yRMpisDOENZ1plHt797w3ZAuYqVsFG4QS1tg7moEFC8wPLlNIojEruT8vq1b/UsPPg1m5+Shnq92y2ZVjibEcHUdGBB0a0n9us+KGFHw+0vN/o2MLjk3mtavRY/K7IEDDTA0L2RLLBApcuBtjmT7+ewC6wHAiCB2qnQE1rDm51YmMX5nVptsviwv36DwcbWkrBb2kxh1g9DGxs+0vyC0olLCMGjdCnjKMSQO5ExTphUeH+bGHQpy8p3I4ek3Qb+8vHjqTCpBY6VoyZM3BTXvi4KDJgxMQYxepEVpdBVkqBP2hBaUZrLRQRai0yji0jaFX+IN8x7KofspZRRvG3JJjUWHWcptE0sqkSmh4BuoZzGMFdq8rSqQ1PgFNeugy1a6S3xm7mvaUwAlZ9SqIa+CD0LudUSQ6kCtxCYCTCWLKcYU2HhOldBNIsCkQVBVRQw9VKdeu1i1raHXW/VyvMMndS0RnlFy7Dt5oSxOMuq+SJeQffEcXapxdnPJ+Ivf332F3JndSXR5H37AlP3HSsUZ8bEm/psafnWc77/ac3GgDjCqcwCp774B4ySkDldOcEoFQxhHi0hbpVJVtlNBFGi2KnYqR6ZU33/UyuEFO/Q0q1nad+1/ierDFmmrgKlkMghGgsiBMObjcUphMZQoRKsR84lOEkMLws3Uv24bqSCGSiMvUQPVTh5wllsKqpUXYG+Mf5QVg3sWuxaj8i1fvTHK6K4MNg+N9belG/d60Xf5J0ywrH4UIElKJnQqDRVMPCkWQkKg3fWvWrmscei3yKYnYqd6nE51bOh0rZwIfISNDU519iat+lMkUb6lPAbO/HtTr3MtNjNy2j7LU1sNXSKo3atyqF44jlxKSmXhUyQDsRyL4xOYWx9mYUrYirvx/+ouEHGc2hdyu6gjcCG/e2vaFgV0rgNUZvYTbiFmsOUmgZnYn9/jWUixXkpg69EFssl9AF5wTKSJPGIBGgY10Na+6IedSqbbLvz/Zu3cmzajYzEd2vBpn4wvAcZX6/Zfiwti97QWTttPDbZDjw/t9vCzdIKajeDhFhAGCHlzPuSeW2Y14Z5p8vjWhTmnS6MZoxmvNNlr0CN9yUzvjG+PQx8Y2DjaI3RjNFsP9CMo7WHui+ZZF4NkHutwyciLWaRAkeOtCHtI7sCFK/EdNXiT8Hzcqm1OHvxRljhY18Cbs8lCjbJluasCHfAg4OaqqwKNWZqkOGZ4ZmDzUeHyxxsMpoxmnGwuVegxtQg4xvj28PANwY2jtYYzRjN9gPNOFp7mNTgLnXvfC3rO4hWxbv6+IJu1e1uwWYXLW09pA+S7HbV5kPG8RYj36ZtknqlnQ6dmXiPhrdKmgRrN0ua19+6XdKJxy9WkxwsbSzezkQCVYvyWHWE03ENmHIJ9QgPUMCKHoHQQIJY/L7VHoSQhtChK3Rzn5tF/ycCFrowE9C9di9mt5wn9Cmp2a98QBuFDchb/MxiAQqMUjIO/ZSpWh4uebjk4P/xjZMc/DOaMZpx8L9XoMZULeMb49vDwDcGNo7WGM0YzfYDzTha22eq9pxIwPVkbZsktDwkWieeBlFUdKigRlO1uz6d99qzA+7O3iLx2LYb92CYucXXa3lbYjFb12utO2S1W8BvafRH1Tp1wF3C1pxASeXUh0REdNjCRm64Tu8spKbFbUOZZOWBjgc6Dtsf1wjHYTujGaMZh+17BWpMsjK+Mb49DHxjYONojdGM0Ww/0Iyjtf0hWev7bgboVryfxl2SsXmHbKFycFCV1tduwS/oV+wty2VG6TIL7m2nrNQ6wjN+2zbUfjrMutZp1lCvNSXsXbW18ymj3fzMjfL4xOMTR9uPbWDiaJvRjNGMo+29AjXmRhnfGN8eBr4xsHG0xmjGaLYfaMbR2sPkRtvXYw803F0/r9Jwks0mZVWk9JR+JTX7mVUlvTZXNF90639G8sXlYtWwgAJLzsFS7Q3PUN5YmKNL6Xv5Y3EovpxDGpN57NXgy7F3NRR+zq4XLTlUCLKXRbM3czwszjZho8RpVXpXUUuRS+hk2WxEFYeHotIVnqt64IhdZJHtV/10MZNcuT2mdscp3qmEfRqllXL22dz/TBdcubQeZ9ptSUv3rYYU4Pp4Cxtdug5voRXhXJkrsOgid7qF3ev2HS87X7vz1smni+KHZgY9+5oWYJcTrWQRLODxZRbPCplMZJwv8K6pfJGVGeifsCBUEzrFNi/wPrPWPddVFIM2kaH3QyhXq4sb/9L6W+o1FH6tu7R9CxnuEkLvUX3He/1kCpVWMm0niuPdZlz3XfbFzYNSBVSH/vW6cfL7IHLtdkt8vdzTWK1vKZ3FnxNJ3dLBOHfXe+36eixe46XuzeZyXEcRToIw6hzf6SuANn579Wwuet+8RIXCJGCeBU63RvOkXhB66u2IX3fEyxYhtq27F2LTZXcbYYbv0OMJCE9AeALCdMoDmXkwncJoxmjGdMpegRovfjG+Mb49DHxjYONojdGM0Ww/0IyjtYe6+BWYA0/6uMB36DE8Mzz//4BnxmUONhnNGM32A8042GRqkPGN8W1f8Y2BjaM1RjNGs/1AM47WHiY1eIctmb6D3Osdeu7Guzvvl+TL9PgyPR43edx8xOMmD5g8C2A0YzTbDzTjWQBztoxvjG/7im8MbBytMZoxmu0HmnG0ts+c7b/8Mj1z+d2dSVy+U4+5Vh7veLzj6P1RDXQcvTOaMZpx9L5XoMZcK+Mb49vDwDcGNo7WGM0YzfYDzTha2x+u9c+7U8+7/u7OvCvfr8djFY9VPFZx5M2RN6MZo9m+ohlH3syTMr4xvu0rvjGwcbTGaMZoth9oxtHaw+RJ+X49vl/P63a+X6/hnrtN5/v1/nxV8P16D/Z+PX80ckXe9JRE2OTOL4EYzUEqSHP1pe335jwZPNkEajUBsJXG4NvSvFOwTzEJwgGJxRaBpqg5TgpGRFDSfLU58nQph5e0BgXeF/cxL5QKV60Yx3s0vHBnEgyt2vWq+g3RimdUb6igcyv03Mns1swk6xduJgVaxaAYs5y4iOYLBc0AHU3lNIqjcmVHKkio0Hl0qfLxfUyHwUvyduhYP1n3QQq831KpGNElOVpyVfjVsNDWLeTEmKfGATrMkhQcAh3LCR6LX1AvhVjKuFL4KkAHV15yM8JiAwYLs218Nv7px17ZVVLF4LRL1db6Alx8kcFUDb+PSasgVpX2avTKyCTvQrF3M6SPWX660Ywg0WEunti6PF1rT3T+kW6+v9EJoCtalVZN6KNNOBUMt/7vfxP3ZmVfe1b2dYOVfd3eykjacW807cp/T6ZgGp87BRhc1FCM+Qiq6VxXHdAmjLvw9/Nvm9VQ7drd/M+tuvnrus4FowQdI1zUCezM/iuMU1DTOPqqYgsaertrXvtMj0XYO09PGbyHwbsHVQziDOL/MhDvoSSD+SMFc39O0Sv6BiH8ylFOw2p0LjQ4DZFXUVIlIvXUa1RLsxuZohIbysfy4M08xTA7BvCCrDJ73FZZVdhTIkUeg/M1VUSRaQbIeBUAwrZmWu5TbEOQkw0XKodSsJxJrqDTytWtIPF8jdFMVXmpoIKH34+fEd0E/x+L00xH5D4Wm6iE6H+VSNWl0wLSyaEg0k7R6ZPmHMs8V7JwyqBK60zMZHHgAK47i6Q+jRZZFlJny/irkFMkpUxZeRToseMht2CiGoKp8503eob4AeDij0qlATQK6l0YVg2vL7qMYsM4+qTerCpM04zBCfNdNzKM2EHYOMqGvWbnyAabUIwraMf5ew0AF33TPjXTXEue3YGXe6jcW6vLBro0ryfxQ21cxzrdrguSORnG+i2q1PRdj53y3K8sKnXj/R7QepfLG+S4ukRYnz8zltCfqRBTt70ZelrpkFtkhK/qc2MNCG8/eNSW1xHrS+z6rkfjtEg07/RaIZcyiuU0VsPnVUzAP2fR/FaK6ISSVIVh1+kYCCyl0mozI2TKWGRRsKVPyarMRp5uDkYpZOoxiUZiVyH/7SDX1k/Q+RRGEq3ogfAEmhHYu9uIV1yAX7aUAwgYQRCOyw6IYFNcxI3j7BJPH35bgnYB9kONEI9Z3jbHg4ifKR0iZCBzE16BHlp7uMeDPt7acm40sLY7hhpZN+ygbu8BNhh1NxYnMCbTkcg6VwHGVnadBe0NW45CpH0ZBfSuHfM2Kh9jSY31mBF7Yha5b1sGI7L5YDSrUtLUpJT4Nl+VC/gRg/H21768Jjsde3vpxSLDYdCtr/vk8cuVsNo8aB0PTWOSLHGyUjbNdgIweAFNTjF2lqHMMXZpuQyuHpGjU5t/pmxjbJddOmqqQKdW2yOlcdELrewf5x/eO+Z6UJKvmmGJMGrGmbmmEAT+l8vwtzon9uffIVSYD5fgaXu4AIgucggwoL6nlFToFQzeV7hEJ0VsEcaVZrXdnF/ulgQnU7WQy8hQCGvdPEcnIkfPYxmoW7vfABAhj+31bKmKAmJ46h7b2aJ11rhZ7PPP/HlpqzU2ZYMKXuTGk1FIbs8d754DZE1g1zJss6CQM/OXXi+mWbBcUwkTYHUqhoE2hI5xZbVg9y18d31d24M3DtzcfEezqzSM8Cx1L8ir+9MfNWAyjLtPprhuoQHNAfpaEzrbfRuHuBZmm5GGYgOzKOLhxtq+fwOjagKRs0OnZhCaV6SqJwYaKoC8p9AGs5fkPlYWfOia6GChEgzq5qY+fXttZrNehqE5rQMcM110k+Z2KY2m/eegOv/nDqHenxBVDU0dCQRN7d0uI7M6Zu0NgxHQDugMRloEOYtD56/+eWCUF+FoWiJVJr7kq1CCXoMv9cj6DdN1X6FekPEPeHxOT93XawZEh9isTnDkNfLQ2mjdTjeHMp+v/dst0ybZ3iqdzDYza56Bsqfp7BFa4cv3P4v5IKzcbpPfZF+oohaZTOXubFS2vrcb1Ddx+BsQhbGSsfIhYGVtTP9+G2XU3GvU9MnyTgnwJs7muFayFXm4flliHZ/+C0xnEpk66t7eTIRLgDlRzVn6FI3OcKsC6uIt3CCzYHb87Lb12Z9Xz1xXIl8wKN7UxrD1rk5mJj1tON8d4YQgyOwbc3uaJlNZBosWLYoPTmBW3uzIOkPqWJemyxqBlkRMFOAf/BxBNal3ZLmAX0fL50dROoPpThqoIyrmEPdhHTYrErgbTBVLok8/2d3+o0VZ5sdHR9A5hwsQqMLDGLmEQ41cwhgy03irVVAVES4yQMZXrp8/XeA7BBpb5ZdZuHptLpxCC6kV4MFTRxMozxxZMMJNsN7Wte4O+quy7d42V33zVy+fl8geuVrvR+1tzh/M5R+KMFCr1s7ozgZof/fsp+uhgj0Ovkuwe1tQm/TdDY0ecX4xuBXO36RmXdVuNuisjD0bXk167tZcPrmiLpqVh0/XtzWg7Yh1E1r0vdl63qZmWyyqIwiHCDjHs61lZOqZ++As3KqjPUK2w6ibLiYa1HuGz8HNMnpmXdgn3879IQOkorcZNSyf91dDbfhCvEZrrLFUqi+ZmDvLnOjPqej8I91KODSbvnLRtJ4HUiI8y7kwarAnPUMfzCxZA0JloT7j6lWcFUQG9oUTWxgraAxSMqZW79C5DTGGSATwjjqxBvGm3rdptoK2Nm/Sxyjg/FjberNnVa8M1lDtYrjeKkzjjPdcjjUMBJujPJYRLUbYr6IMzn4y/VgjLQ7ig1gL1o14ijmur3Gx9LcivrnBx4CUBQIp/LmURYSLFgSlkEHJEDAQveurWpmhHht6+NGEC7Qca4w7tpZyRJ6AoGVyvAgClZe3pr3wxpDTD+cfsQmZ2d6EioCnhbxEyhL+ezxChsl8dIEJ6Nn1KJbpvKJPyown3uA//wctDw9e +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate chat completions for a batch of messages using the specified model. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams"},"tools":{"type":"array","items":{"type":"object","properties":{"tool_name":{"oneOf":[{"type":"string","enum":["brave_search","wolfram_alpha","photogen","code_interpreter"],"title":"BuiltinTool"},{"type":"string"}]},"description":{"type":"string"},"parameters":{"type":"object","additionalProperties":{"type":"object","properties":{"param_type":{"type":"string"},"description":{"type":"string"},"required":{"type":"boolean","default":true},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}},"additionalProperties":false,"required":["param_type"],"title":"ToolParamDefinition"}}},"additionalProperties":false,"required":["tool_name"],"title":"ToolDefinition"},"description":"(Optional) List of tool definitions available to the model."},"tool_config":{"description":"(Optional) Configuration for tool use.","type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig"},"response_format":{"description":"(Optional) Grammar specification for guided (structured) decoding.","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"},"logprobs":{"type":"object","properties":{"top_k":{"type":"integer","default":0,"description":"How many tokens (for each position) to return log probabilities for."}},"additionalProperties":false,"description":"(Optional) If specified, log probabilities for each token position will be returned."}},"additionalProperties":false,"required":["model_id","messages_batch"],"title":"BatchChatCompletionRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-completions-for-a-batch-of-content-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-completions-for-a-batch-of-content-using-the-specified-model.api.mdx new file mode 100644 index 0000000..6691ce7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-completions-for-a-batch-of-content-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-completions-for-a-batch-of-content-using-the-specified-model +title: "Generate completions for a batch of content using the specified model." +description: "Generate completions for a batch of content using the specified model." +sidebar_label: "Generate completions for a batch of content using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztXOtv2zgS/1cIf7kWSJy0aG+3OdwB6Tu7fQRJ+uGQBllaom1uKFFLUkl8gf/3mxlSEm3JsfO4RxcKijqmyHnxN0NyNMzNwAhb6NwKO9i7GTzf3cWPVNjEyMJJnQ/2BvvsNXfJ9I3OCiWw7SgMYVfSTZmbCjYulWJJ3cEOB1uDROdO5A7p8aJQMuH4aOd3i0RvBjaZiozjb25WCGCjR7+LxMHAwuhCGCe9SCPkHXXjxvAZ9JJOZHb98Ew4IxP7UAJRPwvf8wn0W7TSCZgh55lgekwmCePmW4NLrkqBBHQuvo4He6c1KQkWmggDneqmvMxG2HLWSb/MBBBlRLHNqMylWy/ok6/0G1dPGdLEQUgqE9yWRmQwZWysTUTasxvMgQNPU+kHH0ZGGnNlxRYg6Y9SGpGChpXVKuVBGyedQqk+05ODGkQt+fYX2DKZJ6pMRQq/sP3DA1bjdYgqr1Ttk7RBLZp/xq3VieQOCNWojckhsQix6+d6InJhiF6De+bEtUNK1uni3IBFPdaXqQmYSLSSyNNzPT53pcmplb5lwlo+Qbvo0tFjfSFyO2gh4ojos6vprBIGRUDWBUwCSKH0BLA8egD4Kwrno5kXo2tINyhaiF4W/61MaJSZsQwCBFiGeU3hA6dHGgbsGfLnI6kkkb0TCNvSRzg8wYZPenJIFloW7tMyZ/KJZs69pB0QrBDYlr2Twt0UqgC6CLBIqXaMbmlWB++x0RnjMXqRlbCuQ6vKmxY6Bz/cYhDXSDnBkym4aVE6dFb0MB+876SjHxLptGLxWasYEepWb44Sveha69CzQz92xS3oAVFIpgy0y7gCJTOQ8tHWNuu4K2NfqRaEZd0+npwcMt8bNEopWgX7rItV74zBaI6dt8AqdqqNY7bMMnS8sIQI6nM1lTh/XmkjeVgIeB6eEx+ChuNSbcrZ90bWSuegG5uWGc+3AbspHynBojEL4iAnmYPOebKBllHk32ffjj4FZRIQfgRrnCWHg5mF0eISFjZtcGXB+fRRk48g2BJzW4hEjmH50UlSGiOA/ZJcd0FzmOFqtmrjRfgmM60wnlnAdFizhuxIl5OpmsH0KKWvLDt6/4b99PPuT+S54poj5GN8AdRjf0rZkYd4I8+GyA+e8/xVt+ckSuL2YQrDLf7itIbB+awijFQBCBN5KeAz02VOUcXJTPRO1TvVD+ZUz1/FuwnNPiPSg2fZ2LX+qUvwCbCQuE6EwL0sMsVtAFMyk27IDhUs5YI5QA+fcHASBQ9NtVK9XLVSWWEuYfLBnOhKAsyDwCpzcQ32xk2GCGboXat3rR/ItV7G69UBIht31Mce7Q3/4F77bchXxkiH7GsJSBA8o1VpJGDhybUDg8Gz4F6pGPNSuY7Exy2Ee6fqnerHcqrdLm4buBB5CUKNTyxqc5CPBVmknZH4EE63cSrQwyecxugISV4D00En/sbYwDgDVCvMHRbc8EyA3yPHM29NWFVf63SGKj2S5xG7c5lulvCRKbBEOU2decPxCCmA1pCSab4lK2EPPcLt9ATOzbQsU87pk+IZZ8eOJxfgULBWX4L1Cf+XkhPFHSJgYUVPCw1xYBilpc7XZkJb+cWgDaYXW89W2MQ/bdsDhIDzAgSnzOem6qAZtSxmd/BbJnPuQtSorFYBAMUesn11xWeWffdkvg/I44ni2ukrjepKVqPjB1ZECE8wKXecHoS8hPd57IZP9qjfzg2ym/9tBDuxv77YusEn8yH7ogHPbgq9cTxstFTqN3NK5BOYVNrBUcJ7nbRyM5whG/+YEQZ8YsxvGYXVpUnumItFzlGoAfoUwEG9FRON0/MO9pQptuwNvEFasvpm2nxSepZsTYbG419A3jo5F0keEA0YjvFigXowiF+JSIO7WIB0rHAVmYL4vfEKHwAcO3LRXq0Ys3G2/p5ORGni2Ifqhge5EFLxHkT01qMNelW07mVQYhPnVOH77eZ0EctgTVxFYk1R7mDO2RdYBdA8nl1IE2OHDUNEH836aNZHszXRbDla9EHt4UFtPl86wCoBXpbGdKJlZP1Oro9vfXzr41u/W/ufB7Z+t9ZHsz6a9bu1P1VQ22S3Nj+7rU+7ZOYkMhK40GRFapQyfBaTtWCec8p72nZkiTLbh3VqFMkiB6OVz6MGKghY4DSZrQ8MVc8Vrxe7CD7W1nRihEhnCxCMmmJRPod06vfQ4fsANQ/p2BmoLm1b1HuAKZrgD8ToOBA9rmi28t3Urc3cx2wrFBjG0txM5WSKlRVNRdjMV4Ax6Eh1U9aJYvgYuxVdnBeLnl23dFuVnm9oVHS4rEAklyYWIxT5LTN54+FpmeF5qrNcWCoBqQgP2Ue0i/GlnvgoMfQivOlOLEmBTmZBx93hq5ct3mVWKu4kLJcLVp/CSjbVsJLii4m8TJQobSTRW0+TvAvJPgxIJ7o4XAsj6LRdsCdBlqcr8USvdGzzGshmXClElRW+jsaXT9K7hqRb+3/8nT0ayi5aKLtYg7KLzVFG1Na/E/1CUPDKF1H9KMoJbAxNcjO5lThgTay4cOzZ/RYdkm5xmn/daJovVk0ugBILlV1kj7DxumCZBkmVvBBq1hSe3msjHiLsg3cPffDuDt6tUNUH8T6I/8eCeCtK9sH8Bw3m8QGkxXqOIfy6ugbSbcbKhbpOITBYZmXG8si83rQoa6iOaS4mhDRFc07xpQ0+4CVVze5MlwZBmRUQFxU4XyMiksy1C9WOTaXEX6w/EF27kL8gDBtRABfkc14ImDQ3uzVIPFsBmpFwV1jctf18uEvFFfA5ZIfaSnKfEJuIg/yXYLm4qqyAp/2U0UUa4TCOwX8zBmum4KYyBgltNRtzs1UFuKp8pdKN5lROtaYqI8fVRSgo8rwKmfj7Inh147YqjuXajSV9v6FnsBcQLv4osQrHotxG1BearqRSdAmovhQEco5L41XzgGO+2BvvHeEEoXI0DGcNp4gHtYlMxWh416KnAN+zNrQP/TGXZt/XNp37rNitJ98PMCiDOamKs3zRFt2qKSUmap4AzzLBRS19ylLhk0iPcWat7mthgdF5qC7agg0NydMug2riZDSgK1qGVGAIRFU4XuTSBOW4HUwXf+1YAlbdymrZIi+Vik8EI60hnOZdlxK32uhcwnH7ZDE/60yN/HL89Qvz0lfpRV/kBss6pjnBGGgdsNmQHeClgcOZm8J0H7/9dcsbD/7pscNNGPutmKUc7Jr8VpWQ3WchiA0agfYXaD6m1uqG0XuP1Y590lhOStPgMlJyO2C01rO5sveQ89i9kRm6bI7Kiubint+3gbFH+fgHROHrL+/ZpDOs3I7Je+ELTbRwTCG+dwZVkPd2QN3rdLgmovSxso+V/w+xsgbTfx+jfdT8U0fN+Bi2xGHeeZ1+Fe5uOfCuOql91Ff+rmQ4ljypb1QXdIjR+dNwewF27d23yzdw5ZWb64NxU4e/teLyOknjz4GVTP7cMGpOE3cMJ3UJ/nKZezTlrQvg/u4qzVZDC7b+Ak/KcIbT8HUAEtLEcDeFbzuXz3ZkdXFhh1hsNydc6Oev0llyLnq5P5g6V+zt7MCUbE+Blki3Fdbvb1us3x/CYFplrUhKI/HQCgPfVrN7eobPMLwcNTcX3jV3MpqbB62X31WZ/2n9pIFRqHaoaxRCFQK93o8PjPR6vW6Yb50+nERNIbwn9q936w5nZ51vMeP3i2F8yBMvJTV2uxMBz6rjcm2Ns+5DYyC+GNAXV/35sgt7J92d0x2jsaa2ALn4psZxHOGAKsLEu8/ls44iCr/axpeIfBAK781jypR/AXkuISTZ7zlb+uFVSrC6DWIpseH/NEaijTcDnnEZBhgz5pgNIKLciO94jFfa0KWnNvERphyVAGWwxsBL9Zkun9ASi94D0Qhtkvvt4qPd7VkwWHSX5/E4REjdKRSXOf1xHA90Hw9O/dzVEQHXmeWYAIhGv8fONzeYJPpm1HyOzeDRZuYvJtEduRFi5hSRORU8hTiPseBCzELuHOTdPvGLWfhTQO1bS+hgfsR+kojC3dr3LApzh1+PT1D6cDMKbQCthl9BI/6/N8D8B1mbUE9tNwPF80lJIWDgaeLPvwEwwRcl +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate completions for a batch of content using the specified model. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams"},"response_format":{"description":"(Optional) Grammar specification for guided (structured) decoding.","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"},"logprobs":{"type":"object","properties":{"top_k":{"type":"integer","default":0,"description":"How many tokens (for each position) to return log probabilities for."}},"additionalProperties":false,"description":"(Optional) If specified, log probabilities for each token position will be returned."}},"additionalProperties":false,"required":["model_id","content_batch"],"title":"BatchCompletionRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-embeddings-for-content-pieces-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-embeddings-for-content-pieces-using-the-specified-model.api.mdx new file mode 100644 index 0000000..c481230 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-embeddings-for-content-pieces-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-embeddings-for-content-pieces-using-the-specified-model +title: "Generate embeddings for content pieces using the specified model." +description: "Generate embeddings for content pieces using the specified model." +sidebar_label: "Generate embeddings for content pieces using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztWdtuGzcQ/RViX5oCtuQGvToPhdu4bYAkNmznoXCMgNqltEx2yS3JlS0I+veeGXKl1cWxk+ahAeQHS+JlOHN4ZjhDzjOnfGONVz47nmdPj47oo1A+d7oJ2prsODsxQjonZ8KOhapHqii0mfgDYY0SY+uEknkpcmuCMmEgTunXcpjQXkhRaR9o9riyMviBuCqVKHStjMcKstKBZQe0rk2sbaGqQ9+oXI91/kzMbCtyaUReqvxD7BW1CrKQQYrW06wht/rhnD/f6WIxyA6ypByZJpum0rkk04bvPdk3zzwE1pK+hVmjYLEdvVd5wMTG2Ua5oCM6K+N7YxkaDNVB1Y9pNy2kuGyxONiA+WUCaQXBFEpYl5CGHkKbpg3/X6jZKAmZgdc674E3lpVXByDbP612qsiOr/to3hxkQYeK4Dldtl4kYmabQHUdDITUhpSZKKOcDKroMXQAfUij73eRmnAhbRTgupUe0E6BTiHA51pWoHUNLb8Yc3yQoe2zQEPsBDTYtO2vq6tzEUfDvEJliyU0y7k+OJi3NfXUOSjPgw9ABF9aF4Rv61q61Z7zmNtSgzg6Gu20NIH9GNsd+3kd5ifgrR67chxNS1fWwDZRtrU0h06BNKMKLFzNWVOHVtIGNpv8EVY+OWsiu74VJ+LNxctkDHF1pMBMMCBY7Cxmq6kCc52CmbSfvHNCjixciBbvyC5snrfOKSy/odensDntcLdbS/D61Gapu8HrorAYO1uzDifnLwbiwraTspphe6rK3npx8cfv4qefj34aEGjqTtZNZEbHL1B9td5vshAXkeIrfR7J/OQ5T3/Z7Tl5peEVosR0T1+CtZhsZp1gkgoiTPRU4bO2reGgFBCI9k61d6qvzKme/rJa7wpMf0VMT57l+671N07NUgIhdZcrVRBs5Gw4l3Au1xpH9nmlJBQKYI+c4OwSFTpdd1L9cN9J5ZWbYvMBJ7mSAjxErNaoO+DNp16CYe9ae9f6ilzrh/559YKYDc3FZWT7av3kXifblO/AKAbirAUTlKz5VBopHDzGBgCGvuRehRrLtgo7K5x7Be+dau9UX5dTHe1a7REuxF5CVJNU4F7DH8eKESH91/X9M5VbvWqL2ZM8RTRa5cqnQrGHM9bkapGK8kY6iZJSOVrsJgKJA/U3W8zImi/kdF1x+jCp6JzVBZYkPV3Hglj6gk1gVaymUzHcIn0G2chfloV07HJqgiKcD+lbHUrxspK1FJdBopKWBif3FHvB3jDVkhdJpTTO96KxiAq8swkAtsIadTYGTA/eLiTbsI33D90Sdg90sXcbNiiGigLhq5aTSPIUVnstfWCfa/pVayNDiisduB1dSLWBOKlu5cyLt1HM24xjAkt8cJdbV+2I6hwa0lIsiGocvsGgDiRf1BGjAg2jnmMeN5zTcotnI+RqP35/MKeexUC8tmB8KDGa5iMVq4qY7lXKTLDRnON54vZD2urH0ZGWid2CecHeZGNSqbxtHXzzkwIUrdwLRpDPIR7m3bPRtD2nyDqJ3uiKgGzpGps5PaWkN2LNQFOBuKTkx/VcF/mCZUi60tqQngCJZxVb8CkIsI0dr3pQ8Hq/R4NfgI5bNp4ks/qczXpu9plOFNRdWPOhZcN/ciGSEj2I5T3MNozqZH0WoLxMD08S+HE4Q2/JhCadM31LSe8E5+w1DguCJy5X40xgSs4fGyL20WwfzfbR7IFothkt9kHtvwe1xWKjxK0UvKzoy1kstjLs7jWoywLJUya7s+70ENStmqoguU6t3SuLJ5SQllz2pHk1dkwjH0VxFfPdkUJU0FzbNUhQvbCmlxnTfJBjiswZilzaOnV4UdOLoUHd4tumoTKUsBl0u/cuuNbEnH7XRirT1rQRBskqfqK0cYGbi+1ipFcNwrSxnnApUtpbji5xGRV3hp8rl+n6LQyP7SiDU72KgNgz7xsy4054Kk0Ioxga2QhUkE0b3i3f1h5R1/cUPePZmy9zpN7a05wfiLMehIgXo5l4JYObWV9+kAnqCKr0H97d55gdnDDDUS1Q2LytySs+BuZfQBDArL8VjhT9pxL7V7CDrg/82i5HFaWf1TWV3/lmdcSvcp/gicv6rVcQ3fNcGF862N9WErD/Cg1QprT4mTWWX0MaGUr8Gk6/G+quzB32XiRBOb6E8lwn8YGclSE0x8OhNLPDElJUcVhRZXfoqbIb5DYmMF6hmsZm8sTnXQS8vqE+Kl4vVmXu6ap2X5WpWweWjzcL3Hizw3k6H9km5NEaK9Lm80XL2DJLEoj9AvUyXYhE4QcZYRCXmX6341SPBOjfpMj+3U5fMqcaKVT4t0Zs/CFeqfj+nYpgz7EFUjTdK7h4ORLfqymWjSXdMLBQ6dRbegmurOObn23hI3rsqhSMoUMvavVq5TtEilqyC5uYZn6JC441rHoXGl9EeNpTYsOwqaQ2ZEdKHCO5r+OOLelNYWDtyZ1ITMPmc8pJ3rhqsaDmSBO+kuGLwRFxBPzFBCUR5ZnYHxSGZOkkObyKx+VUVi3fOGze11CZFGec5LlqwkfH3vS89fzs8gqDR+lOiKxHq5O3aKT/xxkRnyHmxITb5lklzaTlVDuLMunvX0laQdg= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate embeddings for content pieces using the specified model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-open-ai-compatible-embeddings-for-the-given-input-using-the-specified-model.api.mdx b/versioned_docs/version-v0.2.23/api/generate-open-ai-compatible-embeddings-for-the-given-input-using-the-specified-model.api.mdx new file mode 100644 index 0000000..1ed46df --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-open-ai-compatible-embeddings-for-the-given-input-using-the-specified-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-open-ai-compatible-embeddings-for-the-given-input-using-the-specified-model +title: "Generate OpenAI-compatible embeddings for the given input using the specified model." +description: "Generate OpenAI-compatible embeddings for the given input using the specified model." +sidebar_label: "Generate OpenAI-compatible embeddings for the given input using the specified model." +hide_title: true +hide_table_of_contents: true +api: eJztWU1v2zgQ/SuETi1gO2m33W4D7CH92N0ALRIk6WHRBAUtjW22FKmSlBMjyH/fN6RkyR9p40V7KJAcWpuaGQ5n3ryhxjeZI19Z48lnBzfZ0/19/q8gnztVBWVNdpAdGnFckTk8eluOqSiUmfrTRknk1gSpDNZEmJGgpcQoG2T8kExgi7KqtMolW9z77NnsTebzGZWSP4VFRdjIjj9THqBYOVuRCyo51Sx3cj44bJE28HiQaeVZraCJrPXKQv8c53Aw2RJsaCCuZiqfiSultRiTuIhaF1l2C0UZ+o5J5+QC9lSg0v8Ih5dxWvF6dXUn15eqyf/OErtj6HiSHXz87nFMDTWX3d4O1j2/vdzm0XIXMYdn1gnphRQcRWEnYqKtDF48upqREWRyy5KfJtaVMvx5kcXHF9lj0eqNpaffnw2jJBUi7XynepKGPh9XmYKuewdRwN0UJ9nmdJRl/1bwiuW4oExVh3gEDkMm8ZB1pT7pZXgitacBKudrrRwVCG2HhH4Ok1sIXVBBs19rZfSGUbbu46HweKT7vjEa2/RPnC2FbCtymNuyQlWN+/JetDU9ilheMf+uyc5W657lS4RfbwPvZiyjKOImg7hCDmuPvAUrpmTIyUBrlMDGay+ntEkxH3gZwU/p5bXBd2oM38oqfAr2Cxl/z9QnfMfcR7WVpLN3wQapd7MZVTYtcyh2Q9DqgdZ8uRtDMXL/E6tFwl9KeJubu7dasv5GIJbt4J7o/FqTDwDnLfv9bFvP4dA2chFaysylVgWTRSk1wwRn+WEdxgcZ6vtk/J/z8xORpAXTVARNitb36uWtc3A+Cg/Adn5mXRC+LkvpFks6ijKJ21U6tFPSoOqZJE3zPO4TCxut9x6VmnZO0ry1tgZnE7O6lGboSBaSE9TTWXEn8SvObPJ7nPLRcZUw+Fgcig+n75rD5HAefaolCEfQpjkTiFspeyHHFvzLm/uKcjVRubB5XjuHDkBrfu2C+SbDbbaWweuhPYbpjuC5FYCzD4cnRyNxauvpTC+QHq3tlRenf70WL/7YfxF5l65lWSVktPgC1Lv9XslCnCaId/7cE/lN5Tx9ub1ycq1QFWIGdc8fgrVQNovWcGQ+KaZqjt4qS1ub2BOCKumhqB6K6hcrqqcvu/3OgfT3jPSmsny/tP61NWoCEaLrnIgvmbxpvKloVaowEieacK8UAeiRU7zXCI2Hru1Uz+/qVJ7cHMnnWypKiRAeBlZt6BrxDvhGTRgeSuuhtH6h0nre71dHjGx4Ls4S2rv9m/I63IR8G4xiJI5rIIFkGbvSmNB4jA0IGJ415bV8Ed4ygLjT8ENRPRTVr1VU+9t2u0cJxSphqEm8SuM0R2ZCMSLZxnDm7/YF/JvvYYwnPkW6BqbJR+3bUV4TergR3w95nldJJ0sCC/D+lym26LGvbLHgA/6gOtxh/qAK7MdOuhYVzUTCMspGoptRlDWu0wAf189y8pEeOZoqn5r2lQoz8U7LUoqzIPMvEEcnnyM3sTrmSsZN9qKiR78vKguWGKXy4CHCtnFbO0MbfHcA10pujtuOYnYCXfNlPh1hINpRWRyfNfMy5gi2zgFJSx5xaFQQBh0U8Jiy3bwGNPOmJpsDUUnvOVAbdiKgVydxO5EBpyOpNfVfO7M2I4JHI/E6EQUhG8jsclTIZ+sGfyPxJvULz8aWQmnkhRcZ47Hrfeh8zcFulNNZSfdE8hw9jnEdOBs9r0HktS7i5RKdzoANfF1VIHfEnIdMyNtwKT78LeIqXi0TBlNoAVm3I7nWRiFp/TpwVMFR/gpHF7bmW2kxZNODHg3PSFcNOXD0SmtUnN/CLXARzxnluI7zw134tR0kpVJYGyRJ1R8kpdfeyGedheBq4gEkhZnF16yy8dUY1DXDt735kz0bDfGn3lhxkKX7t49FVzsm0lkI1cHeHt4EhjNYoWKouayHnst6BD6MI21PoFYVFlHxTXv9+HjJz5i2TjuCe9sReUNQXYKa0u8WNoqkl8weNPfbnHcUkXEfjCBoQtdno7OmG7YTUj5zAsX8yRZ+5MuDX2mjst/Y+5bjCBc0PAeS/IURa3/gCEpT44bxfIQKrCiHK5BLnbGZowPVE5mTT0alowv+iUhbF9v+pvExTzo0LgKOZ8DJq/ddVTAIcEfhmBh0n5/X3VbC1+tmP2u/psyZG/YqjZfNyAEutr4E+Y8prwn0nO8n/R8Y4lSYwc2CNzfMix+cvr3lZcDWLVKTjrfHMWMJuIYCrnoMOQD+C0Eke53OOTxnd1hc17FFrXdwbl9J4zDPqQrflL3sVfHJ8dk5hMfNLYHPj1Unr7DI/x5k8Yyhpeu4dpNpaaZ1/LEgSzb57z/4eA4x +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate OpenAI-compatible embeddings for the given input using the specified model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/generate-synthetic-data-based-on-input-dialogs-and-apply-filtering.api.mdx b/versioned_docs/version-v0.2.23/api/generate-synthetic-data-based-on-input-dialogs-and-apply-filtering.api.mdx new file mode 100644 index 0000000..f39829e --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/generate-synthetic-data-based-on-input-dialogs-and-apply-filtering.api.mdx @@ -0,0 +1,68 @@ +--- +id: generate-synthetic-data-based-on-input-dialogs-and-apply-filtering +title: "Generate synthetic data based on input dialogs and apply filtering." +description: "Generate synthetic data based on input dialogs and apply filtering." +sidebar_label: "Generate synthetic data based on input dialogs and apply filtering." +hide_title: true +hide_table_of_contents: true +api: eJztXV9z2zYS/yoYvVwyI8tpp722vifnz/Uy0zQZ23m4STwqREISGpJgAVAOx+PvfrsASIEUZdm1byqrm8mMRRJYYBe7v10sQeB6pIUpVWGEGZ1cj7598QL/pMIkWpZWqmJ0MjoLJViiCstlIYsFm8vMCi1SZurCLoWVCUu55czwvMyEYbxImXIEeMaM5VYaKGNG4xESEYXFZnhZZjLhWOr4d4NtXY9MshQ5x1+2LgW0rma/i8RCxVKrUmgrfU/bdqfYblSea81rKC6tyM0QHZ6m0nfsQ4eiKsT7+ejkU1ujqLJsdDNur2dKZYIX8a2iymdCx3eM1SCf+I7vUHQj9OTm8uZm3BP1LyAmpuZsIQqhud0uX7vklpXcGCgCz8N44MgANfzFocVRJPj9FkRfDs/eB915zs4bHkCRZDFXOncKw/hMVdbxHoSFN0FHEmG8+q1FAipeZdaMUN7DTM95ZsQYbOGPSoJWA/N9Bbscj6y0GXb7vHnyGh783DbemMmoz0xrP3Otctfj3qCuGZiwl9wmS1SBZ8BLXtoxawx0zEyitHjObNVVAUfSLqHcUmXpBNhERr8bMuULKIlMClCzK25AoCueSTBVzXKeoXCB+cezURi6KlY9CWQXoCZ9Cf3n4uID86UBZFKBuhuk3denftU3WkPnXeExA/tYKm2ZqfKc6xrFiLIRrszVUoJkpWcaDKSwDBgGVQnPXTtOEwHksru27Etj05kqgDe2rHJeHGnBUz7LBIvqdLqDLckCeC6SO3AZ2cMp+3j2S2Amgc7PBKscCigYWagtVoLloCdbjMWUIpFz0DyVJJXWAprv9eteRuJHuBmtVniRtTgxbRGe3rCM0w9vJ+xMVYtlVsPwZJm6Muzs36/YDz+++GGCQhNfHQTG+gWqvm7vJU/ZmVfxdX/uqPnBcr79adhykkyCVbAlVDf4wyoFlYu6IYxUQREWciXgb66qwoG5lbkgoyKjemJG9e1P6/YuQNPfoaYHyzKxaf1XVWATICHxNREiDREJhi8sk7m0E/YBwgXokAXt4QuIIFkGD3Xjqb7f5qmM0CsYfBAnmpILN2FsqkJ8BXljcCSCGMi0yLSekGl9H/urt6jZOEs699q+bj+Y1+mmyjfCSCfsfQWaIHjuvNJMgOMplAWBiTYQTMWcQ/i7aWK3ESajIqN6Wkb1Yqi1O5iQsxJUNb4wyM2W2R179krlOJk8V6p4jrx1eQklN+Z2M47jAvVlUYL0U8kztfATVLSqej1NnaDpcM1zAdfYlUsvZnC3L1VaI6+PZJKhE7clTTZyAFtIaTVsn9BTiL9PRqCVfiQCBq1vxMJ7V0FUDjr82T3+PEI9lilwKuc1aASYK4AbRxWHaB3m93zhjDQSx0Cw7h82eh2qjSOjkUWSVSmEJeKr9QkrKIgNgHygixsiiDIadxWPf7pdPDJHVmL5rO/EDL2WeAX6x20AqYavhk0cuQk7za54bUCOjsznkQMYR3GnUlQ6G3ARDmdCU44QTpicYuMDiORc8slBDBbDJyeu3PE1NnfzL9T/f343vsYnNxP2q7LCZy6wPsR1Wepjx0wUC7v0AaNBU9jVW7kbMFENsBn/mJUKXA/+AO1yEaowqtKAvPdCO2w5Qjag7/xFNwkZDzQOzxsIYVO8czLyAtnoq7/tYl2MoL2snaBxthk0b1c/uyTfOhrOcnrUg0C843Mc3EcCjsdGryJRuPZeeYbfgjpu8Hga2Ip1diAfeF8jQvvt2FB740EmhFS8BTl6u7UNYaSBpD8jUNdMJE8keLs4bdRkkCY6pphT7HcQZ/0r+BYUj28uBxfiVPL6rhBBaEZoRmi2A836aEGg9nBQu7npzZczAVaWxnQ23jP9iWCW8I3wjfCNojWK1gjNCM3+FmhG0dpfE63dXN5Wpk3ufR1I7kVp8AtMDc6lAEOGHy4DLsPrlKxms5r9kvGc4xKi5Atanlstc3b6MwvEJzGFnNeYg9QiVytHqMlFt1BT2UqLCaUGCZ4JninYfHK4TMEmoRmhGQWbBwVqlBokfCN82w98I2CjaI3QjNDsMNCMorX9TA3ep+9ueeJ6oWBsIEbod2Ep4WbXw2rB9ZJUt/TQfd2TIEoAwZXQxn+zdwfPt2uZpKlNI8NGTaJbw0slfYGtiyX94z+7XLIhz/xXiJ9HE/Z2znLomiwz0SMOzWmBJVfQj3SMBGp3C4gmHMjil6yYawU5FikMaI1mHudm0f5dAhaGUDEY3rAWs9/OM1ysHRb/jt3KbQ/yAT9VxkCAsnDKYZ5TqpbcJblLCv6fnp+k4J/QjNCMgv+DAjVK1RK+Eb7tB74RsFG0RmhGaHYYaEbR2iGnas9dEnB7srabJAx5SNRO3FpBV4lLBqKqhlWfjfXmoMnZY2RvMfHY1ZvmxnDmFh9vzdu6LGazJ4JL2/Ism8p0t958LOQflWhoSuGSo35zOKSJdHw77Y4LuPbV70Jxa25YxxtAooa0afHAKCVZydGRo6Ow/Wl5OArbCc0IzShsPyhQoyQr4Rvh237gGwEbRWuEZoRmh4FmFK0dTpI1pBQH060XSmXNkRW7V8hqUYKBimCx3lxhXHG0Qi5TFiuVPNpKWW6MxA1zuzrUvTucdW3LbEm9tinhf5hu/vUey2Y36lNulPwT+SeKtp+aY6Jom9CM0Iyi7YMCNcqNEr4Rvu0HvhGwUbRGaEZodhhoRtHafuZG8QhkVU614OGIpD7joqhy7Lgo0qmaT22lC3fXXeVt9lNV1j1WX0RhNg9/OnP02dWyXmcBGbZcgqY2p/sWiwnzW5e67+VP2BH77RzK+MqTqAe/nbhv5T0Z/JzdLDt0XCOYvdTrtZmTYXKBhZ0UZ5WNDp3mrOQwyHy9EJUdHbHKVLiv6rhJ7GIWOXzV7w5m4nWzxjSsOMUzlXBMZVGJRj/bbRPYlQSwbMpGOdM+Jx3ZdxjRYPp4pBlqPz4FLtKFsO6EMCQ6xe5Ehrf7bPCuuW5dedvQL5xvHZgZbOjXTINeTo3gOlnC7SuVzTXPpzwrl3jWVLlUVoH8HRakYup2sS01ngQW29HLSmYgTczQDx21jedoc72octCN4RlLlHvus37n08A3aTQn+t16HvjgoeHdc8V3z7geu208gn2PRAHdcf+jYZz+Pohc90Ll9euetdbGmtJ7+fOKu2HpYVxzNH1r+mbC3vBkGS0ux/corKHAvDgnD/oKoIvfUT9fKdwCBEnufkWFxDhgXgDO5h3Ns/aF0PNoRfy2LV7uEGKHvkchtjvsbifM0Bl6NAGhCQhNQCidsiczD0qnEJoRmlE65aBAjV5+Eb4Rvu0HvhGwUbRGaEZodhhoRtHavr78SvyGJ5u4QGfoETwTPP894JlwmYJNQjNCs8NAMwo2KTVI+Eb4dqj4RsBG0RqhGaHZYaAZRWv7mRp8wJLM2EAe9Qy95sS7B6+XpMP06DA98pvkN5+w3ySHSbMAQjNCs8NAM5oFUM6W8I3w7VDxjYCNojVCM0Kzw0AzitYOOWf7fz9Mzx9+9+AkLp2pR7lW8nfk7yh6f1KOjqJ3QjNCM4reDwrUKNdK+Eb4th/4RsBG0RqhGaHZYaAZRWuHk2v9687Ui46/e3Delc7XI19Fvop8FUXeFHkTmhGaHSqaUeRNeVLCN8K3Q8U3AjaK1gjNCM0OA80oWtvPPCmdr0fn60XDTufrrXPPfdbpfL2/XhR0vt7enq8Xe6N3621WhmXUwbl2PxOI2QAoMZqRRQmwhSv1TV1An6xMfKDUILJC5RjNwdoFDvd0XhWJb2M7xBSgnnAJoJiq3I11Of0S/pbN9bS5MnKRK9CJDV92EeKJtnHsN8washp/rH1Er+fGbdNisNtOwLujjM4e351PGOKXaUFqkwj28/BKT4sFyFuA5nhfEu8xg66Hr7jM+CwTbCW5o3jsCBjwnamLpO+nmyk4RLUwo8GB6XzCEiTzGgTzcxDYGRASxnpFWlO1uhIoMmGXCi5HpTLOD3G7hKvj1TfHrZyPUM5Haw05bsYCR1NoVDiHTW4WNFpaW54cH/OiPloCTZEeZSieI4PimSTKzxqNSGAyYGtX8XUTdn66xGcIPKHXL1Vav/Eb8bgJVhAEVvLva5sjH9t3pb3pw1fbRctQq932aKNeVCh8b9I6442ZyWCt+I3wQK86YWEv+otDh0/XQw1Hzr/v2SP/uy7fR/PIY18Om3ljzcGWosBDFnPljCtoW6z156VIwIT8m3eojTrhya2+GZhaAl5L/97bxBUb++tsrI+QGDZ9Mp8L1vsH9i8c8jWWZZwFAhWJX2Jpj6/hOy1gds4TBEQkCqHwZ8TnTKEhW7VJfAYaCJNrYAbR3PfqnTNkF+mhyeTcQWMYksbm+hiFU8iUqSLgb1BjjxUO4NqRmPSldR0p0eOQD+CItnFcZuCfkJeQwfDm/8mP2lYAgGctBIAaoZljpetr7MhHnd3c4G0wYI32DT9XXEvEQ2fhUEHwFM9/BR3/ImrnVB2LRxd+FrfiWeVMCTofVOPY6S+ama9xmiSitLeWvYzQ7cP78wsoPAM4QTZRueGu5lfOb13Bb5xr+TmSWwyC965HGS8WlcsAjTxN/Pc//Yy1Nw== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Generate synthetic data based on input dialogs and apply filtering. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-benchmark-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-benchmark-by-its-id.api.mdx new file mode 100644 index 0000000..ddd8846 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-benchmark-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-benchmark-by-its-id +title: "Get a benchmark by its ID." +description: "Get a benchmark by its ID." +sidebar_label: "Get a benchmark by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETi2Q2GmwRbe5pZtuN0CDBEl6WLRBQEljiy1FqiTlxDD83/uGkmzJdtIssJcA8cX64HBm3rw3FLlIHPnKGk8+OVokhwcH/JeTz5yqgrImOUqOxQcyWVFK92OU7CWZNYFM4HGyqrTKJI8bf/c8eJH4rKBS8lWYVwRzm36nLMCwcrYiF1TjSuWYRE0Uud5YH5wy02QZB88wxN0iPlu7jG5V/vjAB943DzZf7CVk6jI5+pqUNieNe18o0jkuZojWuts8xXUug/TEwfvMsuHtpDZZxGUvSTtUcB2s1e3f7dTZumryLauQ3ETIPAAbWOQ0kbXeftpH/rog0eUvOIE9IfWdnHuxtlmuotyJwOaUpyvYhZ2IAAettQhW1J7ExLr4eOVB0EzqOhaZnW0C4Xs+pXNyDpcqUOl3VmMYzGflA4fRzim6OcWaHJ7jYp7NRV7HQcNwSgqSM9hFOJnnisdJfTGgnjV0PkHtVxam1hqTre5TlJGk6T8CW1JQtfdkldRG+r0HbSTLm63Mz9qwW7iV76UlUA/UdflQAhOpPe1BuD9r5ShnEvfENFREy/8BR3bVsAck+BpU0Bz/hweZedzjx4qhnEuXBgoVlSUQNp6X0mQ0QlKc1h+7ukzD9Z81gRJ30gtlMJPKBeYspeYpKP//uo8PMtR9iipMO43wDcP65/r6QjSjRYaEYktp4Pmd1D46x8XlwdCt8IV1Qfi6BGjzTnwUx9wVKiuEapJ2SpoQsZSmfR/9RA4FqfRTPTej2bW2BrmJokYZ9h2hzqmG8Nc2g3DYk0LD4pL93ter86qh6GtxLL5cfm6TyRB8StxRclawI1jTjEAKR0gzUiJ6lqmtQ3TuK8rA4UzYLKudA79oI67/Iom2wl21VuD16B1hegC8blkUE3TxGMPxxelIXNp6WqAXTazW9s6Ly7//Eu/+PHg3YtDoXpZVw4yOX6B6T04yF5cNxdfxPJH5rXIO3+9WTqYVVCEKmHu+wEIEYzPvJuZZQYSpmhH+S1ub2HiDKulFVC+iemaiOny/9ncNpp8x01tl+b60/rU1NAGE6D4jyhk2FpsMJLQqVRiJCyz0CCiAPXIqIRKNl65bqd4+tFJ5cjMUH3CylAjwMLFqQ/fAO+COWhhepPUirWckrbf99eqUmY3IxVXD9rX/Vl7H25TvwMhH4rwGE0iWcVVKCQuPsfEjNe/ktdoDbW05H5n4RVQvonpeojrY5e0JEooqYarJqedsVtsxzwkMA/6EHbzsbcrSuVD46js94SObSjqJ/R1203HXa3DTP3doNoWKp6lkKHYeQ5yedGVY+0BJpxTYwRrz4Gra21Zgt1le3sSdZmExNJnGk5Xo8igZz95IXRVyzDvI8cqJHy/6cS559xq7UZNK7RjEIoTqaDzGV8B+YT2A3NdalnIfBch+jDJbJuzXE2BVYR4NT7rW8/WG3zEZY7htnT6zvbhie3HVUlK25z7svAFm9mYHVqxgP+Cy7KurP3MoZBDtdt1/M2Ljh3ZB8TOdTF5ZNCePbhA/YZRDH3INPXPea3PnchOZkW8mlY6+GaaYdVF725OnvN3QUKOT0y6qM96x+0jjCjiiUTAmLV0eZdgAhMW6QT9u1ZIj0H0YVxpfX+w5lnTR0uJr0tKCT+xAjP6xG+v6aEBiFJLrz2aLRYrvui9OL5f8GJ+GjguPy9hcU64yaJArz9f5qns8mMiry5bjr8Xv9bAzse6UyPARGR+T8B0uf9B8U43xxKhAj+bTUYTZDDnOMqpCz3hr8RvI69PHawjuF8GnpCQ= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a benchmark by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-dataset-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-dataset-by-its-id.api.mdx new file mode 100644 index 0000000..d42cd64 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-dataset-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-dataset-by-its-id +title: "Get a dataset by its ID." +description: "Get a dataset by its ID." +sidebar_label: "Get a dataset by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWUtvGzcQ/ivsXpIAsuQESdOoJ7dOEwNJbdjOobAMg9qltEy45IbkShYE/fd+Q+5Kq4cfaXpIAOmiJTkk5/HNcIacJ1a40mgnXNKfJy8OD+kvEy61svTS6KSfHLFj7rkTvpt0ktRoL7QnKl6WSqacqHqfHZHOE5fmouD05WelwGQz/CxSj4mlNaWwXsaNZIZF5EgK26J13ko9ThaBeAISewPuTGVTcSOz+wnvGI8dmwOdROiqSPpXSWEyodB2uRQqw8cE3Bp7kw3xnUWxaTg1NPFmVOk0aKWTDIVO84LbL/j2xqj672ZsTVVGeYvSJ9dBZQ4Kay2XiRGv1GZfW+eX4JeZEWvk7zCupnzm2JN6xhM2MpbVDRdUUdnSuHulxbg/8JZLjf5eIZzjY9gDBBOuel8r4Wj7A67dFIapuxuypvt6k9ezuDGx63PR8MSkzgI69JhJ79AEbjKRsQo8gt0o1zbaCGssDjJobiTHlQ0YCwK3NgB/RovTEQR7CG13oaCxTGXlmlWoDRbpb8esDUu1RE65ZkMoYuihYog6AgYYZ5/OT7rsbXfcZQdskOTel67f6xWzqRg66UU3NQVatEiX/EgNkkCo3CjSbY1QRz91k9+H2PbXl515/L+pnXMxSJIF+OdZJolLrs5a6hhx5UQHfv+1klZkhIsgYZQX1vXSKxIYbJMxLqKhNuU+Wkrtc36v6FDl9xrImqlbs1DogIjhfzWRW8tnIIRWC7crCu1WyXwbSbpSqs34EL4tuG53wamG8IdWzzLubDC0rYDFNRnobiRJxxziEBQpNSMhlwC6YvNB0vjkIOmzK7QtmKPvQQLvsoOkg68aDLH7vVDKdNjUWJX9MgBDbG0Wd046z4n6wanXC3b9n+AVbNXC1znajwLYhibAAmgl0RZScwySBWtAzf7mBS1e71ngkCKTgGDdm/eR4seMFJsevQ8Y+4Dx/QGDdm9Waa2A3kJ4TtN+cNtvybpkjTUSbOVH36TyVkK+nlV3Gms0+eUydWsp73pdvbsy2rp/mdEGdslQlCBynTGepgAptZoMlYGCUtAq5n8huJEtIdfLXYUK+QPJhDSWTTmlnJgsM1qm4Ar7FSL7/0oYwN9X7ZhBGe446G+drfeXl2csUiOjzQLsam09dGS8tZaMSsSoAJjLjQW6qwJ1x6zJt0WgmeYyzSkKkNBWwjODghFr43jYJ8AIYVc9dudITVsroyEby6uC6wMrYPahAtRWc9bYoZ2kpgCRPkLKp6dlxOgzdoRz4EMtTH1QIEBlzBtYFrPFRLACzg0xyZ4RGHxoKh82d6VIAeKUmTStrEWJJjb4+hafqC3cWGupvBbag5ruUF5TWccTjng4OsPxfm6qca5mMA/C5NSx87/+ZK9/O3wdTj9xy4syIqPBF6C+2u8PnrHzCPEVP49Efu05L97s9pxUSXgFyzHd0QeqWUzWs2ZhWhVAGMuJwH9hKtCQbmUh9k61d6qfzKlevFntdwmkfySk157l2q71j6ngE9CQuE2FoAsM2tRyL5hC/eG77AxnPRjyQA8f4+RiCoO2Oale3XVSIe+awPhQJ7mSoIQFtqm0uIW+PVqiVsPetfau9RO51qv2eXVCyKYs9SKifbV/7V5H25BvlJF12WkFJAhehFNpKHDwaBOy1Kxxr2WFt3Vrfc/Ce6faO9XP5VSHu3Z7hAsFLyGo8bEjaY6bB4OtW/x3qM34soIezsKl/ckxvfmU3HKUesK6UPTqeMNWk8YaUdISJff5zluvk+PNxwGYchwflFa69rYSnW3Pa+pkuvUDF7kBaTIOFWbYsJ/0Js97zUtIb77ia0G1aog7kfHKkrroVq3f6+G8P8iNg8oOlOIFP4Cq0y90wxbuF52AAqWfhYnHTZC5uqYxgl1gsLbIB5rPLmg+u6jBx+tnIto8KmLyfIduyFfdGmp524/aK4fLs7oydwPNNn4IDCIk5EJnpUEYcqGqxirSIuLYCMQsPMdQUB5xlNtxUW7FQBOYwq2JN9uLD6mwUPA7y8cNVx/p6cwFwNK7EkIC6aQGxz1oWlPBfBWI75tTQ8GLW98rFXKseOkaQlgEwVWt3wbfnaTfAiiMRrYmsnm4Av1k1WJB3Uj4LBkZnyFkDsmiMHkmHX1ny5hwJ9tPz2sEP2MPYX2nGM3lj6Z7SLrtoBY+v4jZupeFa6AcUZceTcFiJDhKU1H61tSt42zNcd69vYQr/QtrO2Jc +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a dataset by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-model-by-its-identifier.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-model-by-its-identifier.api.mdx new file mode 100644 index 0000000..fd768d9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-model-by-its-identifier.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-model-by-its-identifier +title: "Get a model by its identifier." +description: "Get a model by its identifier." +sidebar_label: "Get a model by its identifier." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuELk0Ax06DLbr1zbvtdgMkaJA4h0UaGJQ0lthQpEpSTgzD/31nSEmWbGfjBXoJkFxCyeR8vHlvRHIVGbClVhZsNF5FZ6en9C8FmxhROqFVNI4m7FKnIIfRIEq0cqAczeFlKUXCac7oh6WJq8gmORScRm5ZAi7V8Q9IHC4sjS7BOBHciBSNiLkA05lrnREqw7l977dK/KyAbZawuTbM5cIyDF1XJsEfFZOSF5xZx5OHaO39LXCJmTVzZiL9Zb5cDqxx0HN2iI/zz0zPeybwgTumH5XtuyLTwdauTVBVEY3vooIqg882FyBTHCwQb21maUyOueMWCH6baFo4m1cq8XEMohhUkhfcPODYaS3rf7PM6KoMFStKF937olsseesrhTmvZP9NN8Mp5tbCRYEPGJePfGnZO7/gnUfVD9t5lpItwHGKeR+BeJoKss/lVY9KWsG3OSLRrlCVlGisfY4xKeCq+wqxi33ltnHtvOHG8GX3RR3J+n69ne9ELdkmPNZkseFOgIkSpMGsqekGRymLvSjSTGJLwOro4uKSoU3A6NGdysL742jwLD2C4XYBVdMJJ2mu1/SU1q3Xz8E759LCAFvEz0oYSMliR7p94tdBdKrYy3fb806+ky1G4KDEMXnDTLlik/N2RiasAwyIxHjhhX9Dwh9iKpTMb/u6WGAlyts69sgtLl1wKVICtOASa1VA+us6HDYiV9nOPIFmMw9aP6y/p9MrFmazBNPzog9AvdRJvhhDDKPJqDBmc20cs1WBol42PQb8nMdcJDkTIWkjuHKenAhq+N378bx2XMhDPYfZ5FpqhbmxvCq4OjGA5Y8lsM6aXjjkSWBL4So5IMujb2Ug5jGbsNvrizqZBIOPgVUWWeA0VhZXwwKQIoZ6NNXTV47xWFfOO7clJMjchOkkqYzB/gdbcf0fIdQVbqrVgtchuofpGfCazy6bY5/1MUyuzofsWldZLpdYHin1o2XXf/3JPv5++nFIoMETL8rAjIZfSPWNvz94yq4DxTfxHMj8Wjlnn/YrJ5ECVcFyXE6SRMg1LsbGVxsmq0iETCwA/xe6wjmErSjgTVRvonplojr7tPE3RaZfEtNrZdmutP7RFWoCEYKnBCAl2Ehs3AGTohBuyK5w84EBOWQPzzhtVPFH03ypPjz3pbJgFlh8hJOk5L91WJtKwRPi7fAJahjepPUmrVckrQ/d79U5MZu2zDeB7Rv/tbwmu5RvwEiH7FuFTABe+K9SDPjhUdpvTdNGXu0Oe+dI+x+G30T1JqrXJarTfd4OkJBXCVGNZ5ay8QczS8H3g/0KDusbDl/xkgna7bXHQLoWKrnheOwDY/1JXOFDczsQjoaCzJTc5XuPuZ37lroUwReWNANHDjaYO1PBYFeBzQF+fe8PoLnGqVHm7z6823E0WrwfebN2tGoiW9PViO89IfDKEGS5c+V4NMJv/kmu8aCZnvjLpRN/uTRMdBGRFwsIonBLv/Bz02ju7uk3op4Prq5K54zKbmoC8voehpwHKBbv96BDerU95vKulrqW/TVSfSS33xXb+sPmAH5TDiotNbYii9r3GxZhsOuYQEZ/r0B9ysx5AjYY5Qa+KyKUNl5pu8ZjOlxI1J7hWRNVoJMnbYk4YlsgTGpyvMipHhCrTUt+eWVNCQdPblRK3HNRBL60q5oMdwHrQAccjFuqYvGo5jRltYpx53Zr5HpNr3HzZ6jYOPTtM6bKYulTYWmctv3h2cCPrmsWH7PDWL83keZ+Si2JPFxW9ITDB1h2NefvqXLswnTHimGGnydJAqXrLNz5vPUE9PXLFCX1LyW3t7w= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a model by its identifier. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-paginated-list-of-rows-from-a-dataset.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-paginated-list-of-rows-from-a-dataset.api.mdx new file mode 100644 index 0000000..72f4996 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-paginated-list-of-rows-from-a-dataset.api.mdx @@ -0,0 +1,75 @@ +--- +id: get-a-paginated-list-of-rows-from-a-dataset +title: "Get a paginated list of rows from a dataset." +description: "Get a paginated list of rows from a dataset." +sidebar_label: "Get a paginated list of rows from a dataset." +hide_title: true +hide_table_of_contents: true +api: eJztWGFvEzkQ/SvWfgIpTQo6xJFvvYPjKgGt2qLTiVbI2Z0kBq+9Z3vTRlX++72xvcm2CaU9IZ2QioSabOyZNzPvzax9XTjyjTWefDG+Lp7v7/OfinzpVBOUNcW4OBDHcqaMDFSd5MXDYlCU1gQygdfLptGqlLx+9MXzpuvCl3OqJX8Ky4Zgxk6+UBmwsXG2IRdUclnJ0F8lnZNLLFKBar9rt6wqxY6kPr5hxxo6mhbjT+sdptW6WA3W3yfWapKm/8i09YRc/4kPTplZ/0kC1HuQkawuVqvBrUydzUlo5YOwUxEDEFPrRMDTsnUOyRKNnBGMFXPpP9fWUS/CDt9to3/NCRaiGUdC4j9vzPblQiotJxo/TENcpLzwFNhH63TPfI5sF+SPJ+8iUFmW5D1WJTMcScFB7k75VGpPA/Dnn1Y5qpD6VMtecBeDIqig2f0WhbaQHIgZGXKq5CSltaIjJwDJAIxa20sELbyqG8QM0LUMQ4BkmL/sIi/HxxAJVbmUXiizkFpVAuHWUrMBQP9hZPZBhrZPWwWzM1DsdrB/np0di7RalLaKnMip+l7F3jjHpOLFA07F3LogfFvX0i2ZeEw3imsu56qcC5WCdkqakMps8u/RTyRxAIvu6zmtZtfaGsQm5m0tzZ4jWUUi9vbcgMOelEHMprxHlE+OmkS5p+IgEjQFUwL8hETrQY5gUVnspkUnCZMIwZ7lxLYhOvcNlWoKWtkyqbCkW7geQvFc4a5a6+T1qB7T9I3krQk9dbaOGA6OD4fixLazuV6uGX7yx+/i5a/7L4ecNLqSzPY+v0D1jb/fZCVOEsU3eO7J/Kyc5692K6fUirsWBI2mgg/BWmw2y84wWwURZmpB+Fvb1sTeF1RNj6J6FNVPJqrnrzb+zsD098z0rCzfl9bftoUmkCG6KokqThuLDSMLQ7NWYSiOMcp5bIE9ciYhEo0fXTepXnxrUnlyCxQf6WQpYd5XTKzW0BXyzQORchoepfUorZ9IWi/68+qQmQ3k4jSxfeM/y+tgm/JdMqqhOGrBBJJ1nEoTwuAxNiBh+C3Lq6KpbHXYcZK5w/CjqB5F9XOJan+Xt3tIKKqEqSZnnqN5jVMbjoyHR4z/Jt63FFDizXmsO9w6xhfRS1Gl7cNz89GTx69TfNubSK5O3sm1uOTj6/jc7DHjXfisTEVXYxGnHj/gY2d8Jp7sp91Ph+JwKj7gYD9IS7LLCcGowfohW4sDdyw+xKP85uCdeNE6szbCL757zwb5MU6RWqe1MHNu0ttyrosypW6RiQiX4xuLd9891kc03dF3LP7TsZ3vVRrpZE146uNthsEXVCJn+bPiRqW4No0M853H+cPXHbXzJk7GjJIQ1pVjVxtOB9eC41sdrrsO4euPDKRXvQ4JXpDccgvKYSwmep9d4+jSNlUO2QSUjGwoItFQkAhPpYLdRJhluAWxa659jJETd6PjRJk1Z6LbjOWBXqEZVGtusbjA9lhAFGZcjBbPRjlwZUeou2Mvo+tNJVdYnN74UqnjhU0xD6EZj0Z499ybWw/V7Wkta7mHxJdfh6WtC/bpCdxTYRk3vu4G3qcL/o1bYMSau8M73i9Oeb84zY0wihL+2XnKyOLZjiTx3PA3Oqjs9/S+5XhFg0m4UDBybsStf5L5zfvIVI1F/qACE1+clcP0c0l8VWoDSNZUluSTUYjn3HBjsy52/G3jEz7kaswAByFmVO8xUbWPzbNBHjGeOCeZIQ9pbLfTcr15UXhskP9Dg8wSDHQVRo3G4ap32Zm09ymTuVNfuk+O+sPHca+XQi0sMt5yfc1J/ej0asWPU9tgdVXKM5JNN/gmG35Iy9sZ3ldabvXehdQtr4yX0/cHeWfnu8N511Q3bi/4C14pJ9xlHpSoJye5wz4VD5tYO/F1N/Vm2YfX4e5Ve3XBV+94R0XnZsBpwUFZUtOPbOvl/0aTf/vmDG3/X3Z9zXc= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a paginated list of rows from a dataset. +Uses offset-based pagination where: +- start_index: The starting index (0-based). If None, starts from beginning. +- limit: Number of items to return. If None or -1, returns all items. + +The response includes: +- data: List of items for the current page. +- has_more: Whether there are more items available after this set. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-prompt-by-its-identifier-and-optional-version.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-prompt-by-its-identifier-and-optional-version.api.mdx new file mode 100644 index 0000000..2e707bf --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-prompt-by-its-identifier-and-optional-version.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-prompt-by-its-identifier-and-optional-version +title: "Get a prompt by its identifier and optional version." +description: "Get a prompt by its identifier and optional version." +sidebar_label: "Get a prompt by its identifier and optional version." +hide_title: true +hide_table_of_contents: true +api: eJztWO9v00gQ/VdG/kIrtU1BIKA6nVQ4jkMCUZWCdAKENvYkXrB3ze46bRTlf+fNep04TTgKui9I7ZdGyXh+vHlv1rOLzLFvrPHss5NFdu/4WP4V7HOnm6CtyU6yUzpztm4CwdS2Luej7CDLrQlsglirpql0rsR69NnLI4vM5yXXSj6FecNwYsefOQ94sHG2YRd0F7CJngd2PjhtprDbzOGiZPJzH7im7hEKfBXoUoeSZsppNa6YmkrlXNqqYOeP6F362pNyTNZUc/Jt01gXuKDLkg21HqEowPV5DwKdnr04ypYH2Qw+dFdLykyj4Cm7rdTedZa0lwzIB4X64FkFuntA2uSOa2CFsDDzasb7EqGr45Muflz9W6O/tky6gBc90Ygxsa5WQVwqT3cacfTH/UeHhZ7qcFgqX/55J1bRYzCIoZxTc4TQANNvx15eD/5S+0B20uO+QtuoGniFElXmytCYgSfy0SYiuupSja4EFrfafyp4otpq2O+xtRUrE0tOv01U5fl6Fk86O7gvItcAL3qISA7htKfUL9I+hk/OVl8Dr84uEW6JfFRRaPGuqrMBJ1N0x19b7RjNeb/iwrBnQ2w3Svt4kAUdKqmtk81WN097cHo94UODz9JcYQ0IZBGZXjdsTl/QU9ii4EjwTVAF65eVqhW9CSr/At4upa77u0QsApKaGM28BGe0malKg5GOalUJnbj4/2QNCYTW30A8/1xcnFFnTbktIk8SfD8SxTPnpKlifCCglVA2FF7Xys2Fr0IDjjaXpc5LYYYUjaaZEPkANnW/xziR90Hp6qaRO2sJXVkjui/bWplDx6qI+hg8s5FOlIJBzSa/QZV7r5uOo/t0Sm/PX6ZihooLFp3F0zxjqsEclNmNB4msxrYNMbhvOMfwyMnmeescI/y1vH5GE6nDfbdW4A3oH2H6Dnj9qUMTUDrmIIOXzm07LTGoJ7aq7KWn87+f0sNHxw/jSOYrBd7zkF+g+jreE1VgkEeKr/O5IfOTcu493q2cvNJQBWGwksgUkFs8bOa9Y/EKIkz1DMeKqm1r4sgMuuZbUd2K6jcT1b3H63gXYPorYXpSlh9K61/bQhNAiK9y5kJgE7HJ0VTpWocjOsOhjYQC2KOmCiKRc8v1J9WD751Unh0OXQKcIiWW0xC9aQ1fAW956+EEw620bqX1G0nrwfC8eiHMRub0pmP7On6S1+k25XswiiN63YIJrOp4Ko0ZB4+xcTsoenkN3ravvYH+h+NbUd2K6vcS1fGuaDeQUFSJUE1NvVTTrWtest/M9jkHNDgtX+M5aXnfW+/iyuBlMjWoXzjliqRRDisyNC7eF5nsy3A2XCG1eMdyV+687xiESC3q9z9LUw4SYt2M4Fo0Z0ua/VK/PFglsN5mY3ic6m6+M36/O+8KTntpuHj5Qg51H/Y3M0p82UqpnwLLJYAGPqWFcQafETJgcZKNZndHXTg/WqwAW8KgezHoAG2ddLgMoTkZjfCKclhajy4fVrISH/q4Eue2ziSOZ/Rch3l88K9+Lr7/KL+JUmJ+iUSDlZreJL1EHWWDa6FsdncHaDJe/IbQ1FD6Q8/x4gS1zdBm/8HQtT/MMo47BJuiscDMR57Bi3YYkq7TTiEXBgKom6i8v41Rjj8Y4X+8RQh22/lYdqEKo8KpaZ/VKwzeykeNNcARU0wwSZz5RQVswLNYnyu/6i8xSK7+Rk2F10nJNtJgkajzvutLIg8+naz1hlYLQ8RosRjjtfStq5ZL+brTgFCj0F7mdrH7FmpYwi8rZGcRX3i+IU2sqK1YRfL2N00/meHeedLiPt1wnOxMrb82NPNhXn3Ka3iXH2Fc4uCDtiXR7vfTPOd4B9Y/ufVGsTEGnj+7wGD4BmWx6Uc= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a prompt by its identifier and optional version. + + + + + + + + + + + + +'"},"variables":{"type":"array","items":{"type":"string"},"description":"List of prompt variable names that can be used in the prompt template"},"is_default":{"type":"boolean","default":false,"description":"Boolean indicating whether this version is the default version for this prompt"}},"additionalProperties":false,"required":["version","prompt_id","variables","is_default"],"title":"Prompt","description":"A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-scoring-function-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-scoring-function-by-its-id.api.mdx new file mode 100644 index 0000000..d100aa9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-scoring-function-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-scoring-function-by-its-id +title: "Get a scoring function by its ID." +description: "Get a scoring function by its ID." +sidebar_label: "Get a scoring function by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztW21v2zYQ/iuEPm1A4nTFhm355rXrlqJtgiT9MLSFQUtni60kaiSVxAjy33dHSjL14kau3KFJnS+RKd4Ln3uOPIrSbaBA5zLToIPj2+Dpkyf0LwIdKpEbIbPgOJiyi1AqkS1fZJPgIAhlZiAz1I/neSJCTv2OPmrqfBvoMIaU05VZ5YDicv4RQoOCuZI5KCOcKRGhErEQoLy+2pCd4M52vsIuaob+yUKFMBPR5ztuuO8a2jcOAsiKNDh+F6QyggR/61hAEuHFFXor1Sya43XEDddAzmsHwWxRZKHF5SCYQxbGKVef8NpImZT/Zksli9yNN81N8MFCpg0Z7yqJYMGLZONNPw6XMbAKDUbDOWA8ueYrzTqidy3ZHlxSMJyG1xcpHkWCBHly1oiZzOB0gaDVElmRJKis/j3H8QPP/CaEeY4x9lpqH+oWrhRf+Q2lJ3cf7tBTBaZQ2awKZMeJDQTbFPg6GFWDF4J1kw/8c0G/UpFxJIZFfsKmDvn3pcz7ICBX+4Fb8EQDjePfQiiIiHXWKWSGESYh7y6slktqbVs/44pjsMBZZgt0wdlkVzwpQE96gNsWizJIPhZe02AsnMxYLN5YLUOxQJugRLhDMCoS+2j4bYPhKIXG4vGHUzMUkNLqDgFx6enDsW4ZDIYVGQvFlJQMBcJa3CEMtdgaB69pMBBOZiwSp1bLUCiczR1iYVd7H4m6YTAOJDEWhZeoYygGLy9O3+wQgSJrL+DrlsEYWJGxILwlJUNRsBZ3CEMYczMLsdBJgIzORJYXzRzZ3GMwTL0qxsL2DJU+q3WekMqhIJI/bO0Ps/7sBM3PAjkOw13D94XQfQXU+BK3EjNbJHZR6705fNVqSY9ewEjfJarbCjTrBSMvatBQZ+Q7TviUqK3eoDwqMk57irs0wov2Zg65e/dm31PpXBW6o0F5VDV0XfCOhuXRVdNl8TsamEdSVVfDH43HYyqvg+pp2PdeZZdF8WgkHkm1vaEYHg3Pvgr368lupTwe4O+yMO8Wz+NXve+wVicvKqVWwuqiswv60f9k/R54KziTJJ1xPftYREtoQNm50T5JsA7LRXV2wKqzA5ZXY9L1+UJLWTu21VFKq5uCJdzMUJ22NfCcaxE2CvfqVMlioitQrPjMHcv0EKk5kpP6FInGYnBcr169ZlaYGckKDYxrxplV6jYAzmp5dJTmyBhA5nPTS9umtR9Oc8ekH9mzQhuZMqeCVSqsBfLC2XODqMdEpmFmYfGjWdd9AtXorhftw5zg3GmgAcKNURxLD7LJM32NOCzQJYYkBIUORaw633PpvETr9ryuPivawpF1pPkVardBvgaxjNHObN2UQiRscR+iA0tEO+QJrhtFZo+WwrBAj1fNRKvdelF61Ztt1Kgpzt44at5qP7o0aswDt+a2TjN7ZRFKOsxc0QVhaUNlbQEPY6bk9RfMJU0u97NgU1A8eJDRU/2SZFsZs3k6cmCg4CHXh46LnTzHKWQhlhgM+rXNY5hq7mnl93ru6dwYM/e0lH3duYcE6CB1Rzm6z9BvPkPbER+Qjza0Z5ZaWyakNcIcK79CQjqWNx601C1jUrDS8nVzb0/9/5n69xL9D4rclhS30b6f21/0+LrBs8E7kX2pvC+V9wvxNz8bPahSuZmwW09F+8p5Xzk/+IT9livnkrZbJ+a+gt6nwAOpoP2Hym0LW43FexG/+Tb9QV2CV6+KN9/F7iNkZ3DT7kjqd9hptECHYzgm7OBKUVmYvLChtUP8ue+LBPcm/L8FaMOusWgVGWoREUN9KU9QLXI02NmXCtpwU/j5IlDtsmeB/vvy8oy53hiuyOZkidB95fKfSlE5TJ1x0mA6lsowXaQpV6uqXAfb5zoWyFjhBq0wE407dsjK+6acDCIMmhiwLXCWXW8yncgMx8biIuXZoQIM/DwB5sk03CFLAudOnoXbbQqm7O35q3IwITo/B9qBRJSgyDEl4Ir2BApwmBRPl9B8juxwCZxDiJwNmaQJRwGab/m1TQqUEa6iVYPnMdzCtAG8qgxx5Qn5MD07mbBzWSxjnHMWMknktWbnL56xX3978qudteCG03mbzy+kuj9zROzcUXztz0Dml5nz9Pf+zAkTQUdUMYpre1YlJQpnq0oxaUUiLMUV4P+UZnaLrUhhn1T7pHpgSfX097W9S2T6a2J6mVnaT61/ZIE5gQjBTQgQEWyUbPSoIhGpMBN2lgBHhwyyhy85Jgk9x1DVSvXLppUKizgsnhjCSakECA8Rq8jgBvGmbQyUMOxTa59aDyi1fvHXqxNiNnrOLhzb1/bL9Jp2KV+BEU3YaYFMAJ7aVWkOuPBk0halUZVe9Zas83nqZxTvk2qfVA8rqZ70WRuQQjZLiGp8qWk01Y7M35U23f4LDJGsvTubr5jAEvDk+SQoX1CyW1P7blLmTmbqz40zt00UpC7nJu59TnLyvApKxxbGeQlmEviBMKqAg25aVk8L6eQIPYoldg2W9tNsa/k4OLr66ai0cFhvxo9uG87eYXe3HLsBFYpwjY3Jj4+OsDA4jKVGbA+ThKf8EGMSfpqEMrXnVRoQaWFWVvB5NRu9+0D3iJ/W2eo5NMmzC5JnFyVLLXsD+r5caYfO1U89gFFS6wa9uZ9wvmZDL0CWO3b9PmOtPwwu2ModsiiXOF9pnCBsVSPodT/lGBtRRGgyUwse0kkJKeUK3mfEOqlsOnaVz2kHkoB7KlR69Zo28O6hTI44pu7zgZI0Q+jW+ma9nroHCZdEMXBjjvIEyzP7+rGys5+jyDuHeIck2Hbc5DSGlJhAIre3cyz63qrk7o6asW5URAG8tDPvnOKNhIiEpuuonlo2juWH85LrP7LB6dE7tuqpXkbP9Ow7xvgLLz/BqpOkdx9QIMaJnI4r0F3XZxqGkBtPurNCNtLtrz8vMQH/A0Aye94= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a scoring function by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-shield-by-its-identifier.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-shield-by-its-identifier.api.mdx new file mode 100644 index 0000000..52bc597 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-shield-by-its-identifier.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-shield-by-its-identifier +title: "Get a shield by its identifier." +description: "Get a shield by its identifier." +sidebar_label: "Get a shield by its identifier." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETi3g2GmwRbe+ebvdboEWDZL0sEiDYESOLTYUqZKUE8Pwf+8MJdmy43S9wF4C2BdT0pAz8+Y9fi0zj6FyNmDIxsvs7PSU/xQG6XUVtbPZOJuIy0KjUcNskElnI9rIRlBVRktgo9H3wJbLLMgCS+BWXFRIfV3+HWWkjpV3FfqoGz9a0SB6qtH3bEP02s6yVTKek4m/peBc7SXeavVrwye+Ny92PwwytHWZja+z0ik09BxSgtSYU7TO36qc2goiBOTgg3Tc8XZaW5lAGWQ5WlmU4O+oHZ0z7d/tzLu6avItq5jdJMhCZOedD4VTqM3Oqz7gVwWKLnPBoQ8EmHtYBNF24MzBQxn2IQ1KaR4IzPkW5s7ilyklve5ha2NoqPVzTvEj2P4rgimnGvXerLFdvwHvYdF/0UayulntJvbiS9VE9lK8c3aqZ7VP/BEpG4zog5g6LyIB0KW6eiqjKZiAA+Lvj1p7VFzOHq22udEygaoRdTQc5OV+6CciwBTjonXfK0MBUUiwIkdRB1QiOkFkl3eilcSQQuVgf9snoaaiP2oMUdxDENrOwWglKNcSDKVcovr/1BUixLrPDU3DzhIo22H9fXV1LhprSkNhkkwD0GPNbHd97z0Xio2JnQSX81GEuiRJLISbphJisrkvtCyEbpL2GmxMNSYom+/JT6JKBG0O9dxYs2vjLOUmiroEe+IRFOQGRa/PVjjsSZMgwcoDsuwRdiK+Xnxqk9nhgUfqjXMUpfNIaXI9G15D7urY0LlCScyUwklZe0+zB+7E9V+I3la4q9YavB7BE0xPgNfN+WJKs1SKYXL+cSguXD0rzILKY4y7D+Lir3fize+nb4YMGj5AWTXM6PhFVN/4+wOUuGgovonnQOa3yjl7u1850mhShSioe+AGTbTU2S66gXlUIsJMz5H+S1eTDWOrSzyK6iiqZyaqs7cbf1fE9M/M9FZZoS+tf1xNmiCE8EEiKoaNxQYRhdGljkNxTus5BRSJPTADEomhj75bqV4/tVIF9HMqPsHJUkKCh4lVW3wgvCM9YQvDUVpHaT0jab3ur1cfmdkUubhs2L7x38pr8pjyHRhqKL7UxASEMq1KOdLCY13aeqpOXutt/qPz1C8GPorqKKrnJarTfd4OkFBSCVMNZoGzaQ5kgaPfjvYDxsStdB7LF0Lzfm990ONLic3pMR1vLT0w3/uHQc0jVRCLvYftjWlXjtYd1XWGkX1sgI++xsFjGXYH4xUlQMEUjkyzWbo+SH7H2Wj+atSMG0bLjcsVXzCkOagJv/YMXRFjNR6NaO0/KVwg+E6MgRJOCHZ5N5SuzNhRQAJTx0Xq+Gc34Vzf8DemYIqvrc4n7i8uub+4bIkI7W0GO2/gmL/agxDrNmwxGPqa6o+czsnt0Tt8s2LnR4XEtDlHqypHU1KgOSBtXLSn2cc3pFSEpOD5yk9BYmgGBY/fLBPL+aS4x4PnfMgwpEEPsy6qz3zDExJ5K8KRpgfGpKXIv1NrC4nlZm4+oGvLi4gPcVQZ2n1xDKm4y5YR1w3aLSeoNe5xlirIhWer5TKnbdxXb1Yrfk07Qc8Vp2aaS3MuL9Vf6cBttZ4sngz+xUXL5pfiQPrvTaa7ALILphCYmp+oeYeLbf2lq6CCZmW+76NIG4OJlFjFXtdHy92Wlj68vyJ1/QRrXTc2 +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a shield by its identifier. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-span-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-span-by-its-id.api.mdx new file mode 100644 index 0000000..7e24e0a --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-span-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-span-by-its-id +title: "Get a span by its ID." +description: "Get a span by its ID." +sidebar_label: "Get a span by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWF1v2zYU/SuEnlogsdNgRVe/ZWvXBWuRIHEehrQoKOnaYkuRKkk5MQz/951LSbYc222y9iVAggBRJPJ+nnP4sUgc+coaTz4ZLZLjoyP+k5PPnKqCsiYZJSfispJmkBwkmTWBTOAhsqq0yiQPGX7xPG6R+KygUvJTmFeEmTb9QlnAxMrZilxQjRcPc59V3hvog1NmioGbnq+M+laTUDmcqokiJybWiVCQYBPJ8iAJTmb0k7aiDTwpH82KlLQ1Uy+CZQ+VdJjx+d4xPzuLT1I/F/tdNkajuwOhJo1z/EqRFUrnq/SMLOnHLv+uS2kOHclcppoETxLNiBTDo0Muf2xWL09HFZqPODy78kG68Dmo3Q4ReCnR+CSXgQ7jqLtRjPESRspK3BRk7nhNadpkRCb/GSe96n7X30QZ5QvKY3UzW1aaAuUcgAxwl9ahgeJdoMo8V42D8w3IWkNnk2R0vZphaq1hbvV/aq2mmGJvSJmS679pM+29kc7Jef9FG8ny0/I7qf9D88OZ1DUDSTkvmJcSGaPZ6wRESQGICFLI1NahR5vlvjwnUns6gCR8q5VDuUbXK672qNbCcgMyn/BdBc0ZsFhs9e3kDuRiqMLjj+637UYBngZforMBQuVgf9ulSmOkw4ECA+JGgj0GFVG5AMNKqRlKlP9CyQoy1H3EKJidortbXByPz0UzGm3JKYpUU5kf0fitcywPPPiAi1NYB4moy1K6ubCT2ECKY24KlRWsF5y0UxJSwsqCAjffo58IIMBC39dzM5pds/5BrYpNXenN2QiHPSmDnE12jyx7MD4RVxfv22SyKLyi9pRDeNFZzKYZidI6KKhppIE998FMGWQ1EzbLagdBhYpvxvUgoDcd7rq1Kl4P2bFMe4rXLaNi4mwZYzg5Px2IC1tPCz1He7S2N15c/PWnePX70atB1MJbycLUxxegvvb3h8zFRQPxdTz3RH7LnOPXu5mTacULUIHpTEeU3GKymXeG2SqAMFUzSKssbY0xXNtGk59I9USqx0Sq49drf2Mg/QMjvWWW71PrX1uDE6gQ3WZEOZeNyYbtiNCqVGEgzrHKe94zzoWcYs0VGh9dt1K93LdSeXIzNB/lZCoRysPAqg3dot7YmbTFfaLWE7UeFbVe9terU0Y27zwvG7Sv/bf0OtmGfFeMfCDOaiCBZBlXpZSw8Bgbz015R6+cJrLWYccR9TuGn0j1RKrHRaqjXd7uQaHIEoaanHrOZkyacAh0c45/M953FBhd8bpjLhT2e6dv+Iankg5nO9DYx8Nuc//QP/opnl3JUGyf/5Hu6Zuu+O2lihVTWh89Y13YzbrmwdXowRYDu5Myn4zbINYn0fvGEF02ITzMKcqFIhQWQxNMjnWBt1EynL0Yhq6qw5ijHy668iyH7BEv2lCXfEaOUthUs3bcwSKEajQcYgtyWFiPLh5qLUt5iO5nXweZLRP27wk9VWEeJ77pdO/6E39jJsSwW5C85/nikueLy5YPkSfwz86b2sxe7ChXc+3UJ5LsU7tvORQyCAjiTMHIRyPu/ABNFM8IZPLKQhk9pCjun5SDCLqGGzkf/Fk23YRr1xiVjj7yBYa2LhJ/23jKZx0NKXBy2kX1AcKqfeRQhTpCpbgmLVj24Xsj/8V6Ydg7oYVGoNswrDQ2fOwvNnLRguK6qewKFt1VCcvIqMeciA1+1wEZnWQAsIXFIsWu8srp5ZJfY2PquPN4jNKecpuBg1x5fs5X2rU3nWcXLdifi4fwcme23VWV4dTitRP+w+NXmve1gan6CwK8Q9r/EVFX33iHVmDhwgrN1Wu+nmQZVaE3b2tHsEH/d2/HEIT/AL47DvM= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a span by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-span-tree-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-span-tree-by-its-id.api.mdx new file mode 100644 index 0000000..18bcdc6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-span-tree-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-span-tree-by-its-id +title: "Get a span tree by its ID." +description: "Get a span tree by its ID." +sidebar_label: "Get a span tree by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWNtu2zgQ/RVCT13AjtNii27zljbd3WBbJE1cLBZpEFDS2GJLiSpJJTEM//ueIaVYjpVL0b4USBAktjSc6zlDcpaJJVebypFL9pbJi91d/peTy6yqvTJVspfsi48N2cVpLaupJTppF+wkoyQzlafK8xpZ11plktdMvjheuExcVlAp+ZNf1ARVJv1CmcfC2pqarFfRbC79oJTMc8UKpT7ekL9fm4OjFyrvCTpvVTWH4GZgnyr1rSGhcoSgZoqsmBkrfEGCVSSrUeKtzOgHdQUd+KRcUCtS0qaaO+ENW6ilxYqLR/v87KiOGflN3G0yKg3mRkLNonH8SpEVSuc34VWypIdN/t2UshpbkrlMNQleJKJECvFgkNMfSt+L01INbMEPx6acl9ZfeDVsEI6XEjBiJNA4SN32YoqHUFLW4qqg6pbVlOYxIqryHzHSy+699maqUq6gPGQ3M2WtyVPODkgPc2njh4F6F6BNRUezZO/sZkXVaA11N99TYzSFEHsiZUq2/6SNtPdEWisX/QetJ6vz1T2h/0OL8aXUDQNJWSeY5RIRo9jrAERJXjJxhUxN4zdog7z5xm13kn56IZ41NuI0iAszWysZbVWPEC8ylJiv/MVaY5NzSCmvWYyb02m0ulrdleeZ1I5G6HjfGmVRLmjreNejekuLDcjesvSv8kVr7XYa9yP4fSG9UFWmG7zuAlRVhCAkd5KtAhyoLLhsF6JEM+VsB1WHB9wrwmcnrmB5QN33BR3abS+kwfa+FVn3YgMO6G6EknnbZL6xxDUMjiK+Ffv0+9CGwrVnf8AvcSU5EKBN5QLdq5Sag6L8520uazC2cgpq52DOVp+bTo+73GYmp7ABxBQ91CLfMR5FEB4hKa4wFrBuypKr2eI6YBbNRGUF92IO2ioJ+HPXRp3j+2AnYAM51o+1HKXZNO8t2AmKzZ7dW7PhDltSFWKuskdE2aPvvvh08r4NJgubmmgc5QxUS1hNlyRKA0D0QLrRKCjDlpUJk8UmkNEtv76LxB0VY7VukteDeEjTHcnrTkBiZk0ZfNg/PtwRJ6aZF3qB8mhtrpw4+fOtePXH7qtAXbqW3PT7+ALU1/beyFycRIiv/Xkk8lvmvHg9zJxMK26aBZbz7oqUGyyuFp1i1gogzNUlti1ZmgYynNu43z2R6olUvxKpXrxe25sC6R8Y6S2zXJ9a/5kGnECG6DojyjltTDYc9YRWpfI74hgnKMfn8YWQc2xgQuOl7Xaql3ftVI7sJYqPdDKVCOlhYDUVXSPfOPW1yX2i1hO1filqvezvV4eMbD7Vn0a0r+239NrfhnyXjHxHHDVAAsky7EopYeOpTLiT5h29cprJRvuB6cI9ip9I9USqX4tUu0PWHkGhwBKGmpw7jmZKmnDBtgv2f9Pfv8gzusItky9f6UIoHPoOD3ggV0uLyyu47MI0IQ54erdcxRpq6Yvt+QpCPjzoX8K5jnPy7RCLYmrYyDrtuPehDFsk7AYRq/MojL36jckXLPGT+Lwes1x4cwG0NbbqrYpzD4TrqXQDjg0Fv1bZAhgq+SjdhR9KXsrri5xq5O/hVsJKIa/KphRhTZfcVttD6O6AhIKvL+fxShGwslkGdo58YXIusAnXjlDnvWRy+XziOzxNwvV8smwhsZqwMzzsCH0/oqaxDNfC+3pvMsF5a1xAH+VjrWUpx4B69nUnM2XC5XUEACu/CAsPuiZ/ds7vuKgn6/K/W9NluHxnXX3ONxK9G/rTzISUtzl5z66IU3ZFnLZ9JE5CRgnHEStw+XygKHEU2m9Ast8S+5rDFAfAu1RQ8rkSt37AQgp3K6ry2gAGDi08nDuVxeZhY0/JeVDCGLEzmTG4WKm09JmnKNrY0DC3lad8R9RooVbOO68+YEPSLgCRS4zuzjlpSX5vX9hIQo+F969qAe7p2k9qjSMzWw7oWLbgOos5voEXI4kBhv97/dkahRka44jXLJcpTuKfrF6t+PE3nj8xaFA73g5TLjHglCvHn/MbRtwZxbOTlgtxsPjIPjYYXtdBKo4lTEHxDR+/0qLXScMAtcDOCt6zq/Ht2+jQeMo61qu3Gh0PZOOK/SwDyO+VPe8R+/jodArhtG2mJR9Q9hIrr7gr42/w1IQEhbYXni0TLat5AxjhfdTJP/8DOlgYYQ== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a span tree by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-tool-by-its-name.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-tool-by-its-name.api.mdx new file mode 100644 index 0000000..07923db --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-tool-by-its-name.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-tool-by-its-name +title: "Get a tool by its name." +description: "Get a tool by its name." +sidebar_label: "Get a tool by its name." +hide_title: true +hide_table_of_contents: true +api: eJztWG1v2zYQ/iuEvjQBEjsNVnTNt2zpugLtGiQphiENAko6W2woUiMpJ4bh/747krIp2XkpsGEIEH8xJZHH493z3PFukRmwjVYWbHa0yA4PDuivBFsY0TihVXaUHbMLreUo28sKrRwoR1N400hRcJoy/m5p3iKzRQU1p5GbN4Ardf4dCocLG6MbME6EXUSJQsREgEnmWmeEmmZLP3mGU8w1qqZbU8C1KB+eeM/38GL4YS8D1dbZ0WVW6xIkPttKgCxxMENttbkucxyX3HELpLwtNC28nrSq8CbZy3JQRVVzc4Njh8aJf9dTo9smnLduXHblTWbRYN2sEia8lb0XqakvUFOmJ6w7+R7j8pbPLXtF01/5Q+HAb7P11EOBH09InKuA0TrmF+KjsOE5B6nVlB5IdG/pY5J/b2uu9g3wkucSWPKRdrytuFtvW2p0PDmMG16DA2MT+dwYPkfxwkFtHwePQgmPa/cHzupOvtq1p8L1fejY7pGeILYDo+loj4U1e0wgLaZgdv8rI6437ixp4O9WGEgRkKOlgasUZM60MNzyzwpQpAkgWAvGh04mm2gT/CbUTAeOh5NFsYtMK/gyQQqtNletlDhnQ5nkFZIu9z7YZOoAC8mLCILl1fAYO1/8iMtddhL0YjMuWwjK92wmJkxpx2K0KLMlnoWXpQjrTxN0Tbi0kBr3MsBtAzZ9bZIVqKcTTpLqFDRPV9Ab6n+6dilMhPLKeN25N/1oA0rZJ2EdoWJNooTIvCigcR4a+I1T8NpGpe3H/j/8uXG8xKHHKy1ZdxjGc92uI8qP+TDJN/2kEVPEIKgOnZtErYF7N7x6HLzhiLgFVxhfPYdukFT5nPEp6mHRtUtS/6dtufYCD0iqA/r6lltazaUoGQKj5hLxUUP57yVi67hr05gbA9lmnLq4OGVhNiswa/pEFAzxWIh7bwwxkiZjNmO20sYx29aYPeddYAU/57YSRUWBiA5tBFcuEELF734fjxvHhXzqzmE2bU2pDvlWPRhzV+rQTgJzN1fFE06Zopd9PfsUDxMx0FoEgNPoWVwNM2C1NoQM8qf3XAJv20CBWC2YLorWGLxowECvH4F+9HDnrZXxEiB7M91jvO5yyCZ4ofE6HJ9+HLEz3U4rOUf3SKlvLTv77Vf29ueDtz5qwR2vm4CMDl8I9fV+v/CSnQWIr/V5IvIjcw7fbWdOIQWyglW43NIAyYiL1bwTTFIRCFMxA/yvdat8RHXCx/gXUr2Q6jmR6vBdLxuxz4T0yCybUusv3SIn0EJwVwCUZDYiG3fApKiFG7FTTO6okDOUpDiSROJH02WqN/dlKgtmhs5HcxKVgO6O6JtWwR3a2+ETRDO8UOuFWs+IWm/SfPWRkE130fOA9vX+kV7Hm5DvjFGO2JcWkQC89lkpB0w8WIzQZbTs6JXUVYPb5AOCX0j1QqrnRaqDbbs9gUKeJQQ1PrV0Giq8PlCt5suxvsIfwMUCmuotgRc+qt9H/SKOytzQRQpNu1jiCxLQcFdt9oDwyCppKIUaT7MpuFFa/Md+ywbxurqYuhioQ6Vxajb1vUW/31E2nr0ek1Q7XqxUWlLr0YecoHJryFKVc83ReIypfr/SFq21LyWv+T5aubgZFbr2vRILaDvh5n7hSRdfLq/oGyHOKxed8YnWs3Naz84j7ngsfWnzYIPZ6y1mIZraHmB5SqFUsi+JY+1tvyk2+GFMAH8XB1U2GiOQRcr7e4owGGxMwGCJZvTNNjPhBdgglBv4pghH2niCbQrPqaaQSDmDFXjU6jP1fq3HaoN2xGiwbi4+AKNBd28Vgh9YEkHg4M6NG4mXK9rTO3MR3X8ZrOsBgP9Ha1Siu8jLNGWxyPGK9tXI5ZJe4y3PkHtx6ONkTr5EZ5fC0rhcBYJ7Nd45i7jdZY8AfOsJug6Pot6tb73hEw5vYN4jlm/1VBhoqduPCobvx75flazcyGA9snx4f4H0+QfVqZ2/ +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a tool by its name. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-tool-group-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-tool-group-by-its-id.api.mdx new file mode 100644 index 0000000..0e1c52d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-tool-group-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-tool-group-by-its-id +title: "Get a tool group by its ID." +description: "Get a tool group by its ID." +sidebar_label: "Get a tool group by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuELm2BJE6DLdrm5t10swFaJEjcwyINAkocW2woUSUpJ4bh/75vKMmWY+ejwF4CxBdTEsmZefPe8GOeOPKVLT355HCeHOzv858inzldBW3L5DAZipG15tjZutpLdpLMloHKwP1kVRmdSe43+Om58zzxWU6F5FaYVYThNv1JWcDAytmKXNCNKa0wiR5rcr2+PjhdTpJF7DxFF3cN/2ztMrrW6vGOD3xvXtz/sJNQWRfJ4WVSWEUGzz7XZBQaU3hr3bVK0VYySE/svM8sD7we12UWcdlJUiqzvJDuBu0AhNq/6wkj1cRbVCG5ipB5ALb+WdFY1mbL6z72I3gt7Fh0KOwIaW7lzIs3q0FvOMwiq66pVJXVTWoeB792ehsm92znJL6ffxXNZxGn5kawIuBT51KygHmplOZh0pz1DI2l8bQDiv2qtSPFcLNlIBJ0MGwa82/YHUarjsbkgDCxPboL5DC5aMm3xyFLN/HbQt3uyzyxJZ2O4cNyRFkbg4mWzykgJVn2X4EkKRjae7Nk1vKNdE7O+i9aTxZXi/uhvT2tGs/eieHSS4FA6gJReTG2LmLLyRUNI34L3Z6o1pXR6qCH/FLTW/CPlhvaGRlIRX+8KGQpJ/FpQvDSIQsLdu+PbVVjFCnyqyYfxK30QpdTabQSiLCQBoEWpP6/auKDDHWfDeAqTSIM6279MxqdiaY3yKQologGkqf08MU5Tg93hgyFz60LwtcFSsCMweLEUexzm+ssF7oJ2mlZhphZWbbfo51IjiC1ea7lpjebNrZEbCKvkZFdR1LJ1JDojVlzhy1pFCAJMT1tq0/RKMQmmAzOpyRqH/OPzGI0TUkU1hHC5HzGzAmZ2jpE476iDFzMhM2y2jVaXvfrd6jdZrjL1hK8HqUjTA+A1y1zYoyqHH0Ynp3siXNbT3IzQ3qMsbdenP/9l/j4af9jLDB0J4uqYUbHL1B9Ze9PqcR5Q/GVP89kfqucg8/blZMZDVWIHMM9NyBAlt+sm5hnBREmekr4L2yNPoytLuhVVK+iemGiOvi8ti6Jb8z0Vlm+L61/bQ1NACG6y4gUw8ZiwxoljC502BNnWMHhUAB75ERCJLyALVeqDw+tVJ7cFMkHnCwlbDwUE6su6Q548wpILQyv0nqV1guS1of+enVStrvoi4btK/utvIablO/AUHvitAYTSBZxVUoJC09p42ZTdfJaHmk2jpCPTPwqqldRvSxR7W+z9gwJRZUw1SQfXC9XRzDPAaw7fEwBOV6dA0U6ExrbvpMjvoOppJMFQc0+nmdLPLT3CLFzc+LTPE8lQ771ZH9y1OWhZwRJxdGOLaxQD66mnU0NdufgBVyHK7lF12QS70qizcNkMH0/WLrkB/O+ewu+UolVqImgdgxeHkJ1OBhg9d/NrQeAu8bIQu4C+OxmL7NFwtY8AU4dZnHgUVdyLq/4G5MwOtnm5yuPFxc8Xly0VJTt/Q0bb/CYvt8CESvXr3FY9lXVnznkMoj2uO1/lOLeD2WC4va8u6DxqAJx66Id6o9raKn4boUrlhvLjHwzqXT0o2RqWRc1tzl5yscMAxU6nM1br77xnZaP9K2AIwoEY9Ky5HFmraEwX1XmJ4a1pAh0FwaVwb6Lbcekzls6XDYorwiBh8M1xiJ9nHXuOp+n2MV9d2ax4NfYCDpON5qxlKacWyRfac9ttawVD3r/9rzl8zvxDPJvjaa77SlnzB1pan5C84Zm97UXb35ylGS+3ISfTZdhllEVeoM31ro1LR1/GUFd/wG39Yt1 +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a tool group by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-trace-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-trace-by-its-id.api.mdx new file mode 100644 index 0000000..794fee0 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-trace-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-trace-by-its-id +title: "Get a trace by its ID." +description: "Get a trace by its ID." +sidebar_label: "Get a trace by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETi3g2GmwRbe+eZtuN0CLBolzWKRBQEtjiy1FqiTlxDD83/uGkmI5ttssupcAziWyPJyPN+8NSS8TR760xpNPhsvk5PiY/2XkU6fKoKxJhslIjJ1MqZ/0ktSaQCawjSxLrVLJNoOvng2XiU9zKiQ/hUVJWGonXykNWFg6W5ILqg4T2N+tyjqWPjhlZrDcjH1l1PeKhMoQVU0VOTG1ToScRPSRrHqJszbc+lKa33TIfgT7wSeJpyBdoAwflF8Hi29vgypoVyi4KiTQSTIZ6ChaPY4/xks4KUpxl5NZVyImNJOGQ5DJfifAi8/xSeqXYm+sqTLK55T1hJqK1BalJpSarBBdZpmq1593WjaV2hOgpu+VcrAcXq97+KgDGxjd9JKgguYaIoe2sh01GTkqQUTuiZnFRNusBN1TWrG1KGXIhZ0KKTgPVCZk6qz3oqh0ULAWnHBkpO+jGC7nj12EHnO3Gxd30gtl5lKrTIAIhdSMMWX/H9kBR6h8x07B7YzcFhb/jMfnorZG+VnkW4Per1j93jlmMRv3gI/PrQODq6KQbsGQMaIUbe5yleZC1UU7JU2IAgDr6+9jnBX7D1Lpp0aurTm0tga1ibwqpDlyJDM5QWM6azbS4UjKoGaTPqHKDrVH4uriY1NMiuQnJCrParXoLFbTnERhHYRuas1wZDmxVYjBfUkp1J8Km6aVc4Twj/L6L1JoOtx26wG8DvsjTHvAayewmDpbxBxG52d9cWGrWa4XaI/W9s6Li7/fiTd/Hr/pxyFxL1kfXX6B6ut4f8lMXNQUX+fzROY3yjl5u1s5qVZQhcixnAULyC0Wm0XrmL2CCDM1x8yRha1gw9jWw+ogqoOonpOoTt52tjAw/RMzvVGW70rrX1tBE3PesFKiLJ4cIDbs00KrQoW+ONckPW/BCyFnEiLR+NK1O9XrfTuVJzdH8wEnS4kADxOrMnQPvPmEQg0MB2kdpPWMpPW6u1+dMbORubis2b6O38hrtE35FoysLz5XYALJIu5KE8LGY2w83metvDKaShwTd9xufuL4IKqDqJ6XqI53RXuChKJKmGpy5rmaMWkqgOGC89/M9wPh4tVeWhdC4cB3dso/DpTSSawhxx6WiZF8h+3eExUv51vc9q0Y9Z6dtujXvtHFGQV2vIY5uAqwb4muocpqhWSRQW5hmmBxTArhhslg/moQ2poGMYAfLNvcVnxrjYOnTr1yjFceQjkcDLDhH+XWA7MjrWUhj4B1+q2PC2rC8Tyup06FRVx42k6Z6xv+jnkX02xa8pHXi0teLy4b9kVWIj4Hr8GYv9qBD4vVb9BWdoXU9Rx/vsD4mSs4+WLEoz9MBooncjJZaTGHPIQfTyvKYeS4mokZX8R5SLkpY1U7lY6+GGaTdVFm284nfLPQEJ6TszarTxhj2kfGlsARM4Exaeixl00bACzXc3j/ioYMge7DoNQ4YHHE2MplQ4PrGtsHIvBzpAIehg9EReO432y+XE5wZLtyerXi1zj1OW40HuPcnHBX0fZMeX7OHgbD3uRfXDRcfil+zvmdxTQvwUgmjNQVf8LjN1p0lba6gW2OsYv9hdOrvx6lKZWhs3BrP9uQz4f3YwjqBxVdA6s= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a trace by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-a-vector-database-by-its-identifier.api.mdx b/versioned_docs/version-v0.2.23/api/get-a-vector-database-by-its-identifier.api.mdx new file mode 100644 index 0000000..2748e04 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-a-vector-database-by-its-identifier.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-a-vector-database-by-its-identifier +title: "Get a vector database by its identifier." +description: "Get a vector database by its identifier." +sidebar_label: "Get a vector database by its identifier." +hide_title: true +hide_table_of_contents: true +api: eJztWNtu2zgQ/RVCL20Bx06DLbr1W9p0uwXabZC4BRZpYFDk2GJDkSpJOTEM//vOkJItX9JNgL4EsF+sy1zPnEOJWmQOfGWNB58NF9nJ8TH9SfDCqSooa7Jhdsq+gQjWnb3tZ71MWBPABDLjVaWV4GQ2+OHJdpF5UUDJ6SjMK0Bvm/9Ab3SsnK3ABZUyKYlB1ESB69j64JSZZstoPEMTN8bybO0EjJX8teE999OF7Ru9DExdZsOrrLQSNJ77QoGWeDCLvY5ljseSB+6BivfCkuN4UhsRYellORhRlNzd4HGwVjd/46mzdZX6LauQXUfIPAK2GRomvNa7V7vAj7BmZiesxaDHuL7lc8+erXyesYl1LJ0yKjfnNEpsHMocpKSaU4t7MNjM9g8vY7ZQAFs5s+jMgmW1h26uKRhwcfSbyaQqwXiVyNAkVMiYKQ56O+NZa7ubNmWJjax6HRuscM+Q0YajFwXl+rzDsgnXHnrI8J+1ciBp3B3abXKnYcoubvubw6kGFTSV0Ypjp71vm1NZjTHC6EMkFONGsp81uPm66zUMvo/dUX9/7NPlqKCY6OwDu+WeKTPjWkmGEUquMUkJ8vcJ1gceav+Aqf49Gp2zZM0EQhhVmLD6Pwq+dw6Lj8bIdeYL6wLzdYkqm684Em1uCyUKplLTTnETIqrcNPdjniXFD1w9gPwpc7Km1Noa7I0VdcnNkQMuea6BdXw2yqFMCjXOjXhAl8+/VImrL9gp+3rxqWlGYPE5kM4k6c0BesMMUIEOsE2aZ5wc47mtQ0zuKxBIZsGsELVzuCDBVl2P0UYz4XZaK/A6XI8w3QNe+yBhE1z4Yg2n5x/77MLW00LPcTxa21vPLv56x17/efy6H9eNO15WiRktv5Dq63xvuWQXieLreh7I/EY5J2/2K0dohapgBbp7OsC1G53NvA1MUZEIUzUD/C9tjTaELS4BB1EdRPXERHXyZp1vhEz/TExvlOW70vrX1qgJRAjuBIAk2EhsPADTqlShz8410NMsIHv4lKNINN507ZPq1X1PKg9uhsNHOElKgPAQsWoDd4h3wDNoYDhI6yCtJyStV93n1UdiNlbOLhPb1/kbeZ3uUr4FQ/bZlxqZALyMT6Uc8MFjbHxbla28VtuGnU3aLwIfRHUQ1dMS1fG+bA+QUFQJUY1PPXXT7s081b9Z7wcIOOKtnTPL50zRq99qm0ifPCrucOeJyqaYiyxtQzv70rh9VBS04qHY3cYjCuuA7Wy2M+O0pxAo3XocwdXQ2xXnauOLTWFdhUXTbBq/U8QChtlg9nKQEhzJ3A8W3VqX9DkjLk+pndoRqkUI1XAwwNeCo8J6RPZIa17yI5yIuOkLW2aUzQPirMI8Op61a9HVNd0jdsYim8F9In92Sf7ssuEob76dUPIEzuzlHrxI0n6D3Lwrt27kUPDAmo28/27Y1g/XD4jv7WBkZXG18nHHjVGUw4XJJb7GTw60lLkJF+BTUO7guyHOWRfFuBs8p/2HRnk6Pm2r+kwfDXzkdYU44spBmDSUeQTnNiBZrNfvx8Ro6BLgLgwqja9qVFUc96IhylXCf00VPBluEBsHS3wg08WCUn11ermky/GrBQ0f/Wn1zWnqSAupPB3L1fJybyvPLxqmv2CP1cje1pqLyGHqieuazvDwBubbel1eo32BSzp9f8Sik8mpEFCFjvPOs3JDch/ej1CE/wFfGnAx +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get a vector database by its identifier. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-all-training-jobs.api.mdx b/versioned_docs/version-v0.2.23/api/get-all-training-jobs.api.mdx new file mode 100644 index 0000000..081a530 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-all-training-jobs.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-all-training-jobs +title: "Get all training jobs." +description: "Get all training jobs." +sidebar_label: "Get all training jobs." +hide_title: true +hide_table_of_contents: true +api: eJztWNtu20YQ/ZUBnxJAF8dokMZvbpqmKRLEkJWHwjGK4XIkrrPcZXeXsgVB/56ZJWVR8qUO0BcD8gtJc+bM7ZzdpVaZp1A7GyhkJ6vs+OhILgUF5XUdtbPZSXYKn3SIZy7EqUdttZ3/5fIw6fxG2SBTzkayUVyxro1WKK7jqyD+qyyokiqUu7isiRFdfkUqsmPtXU0+6jZ6gbFvhd7jko10pCr8t/eVy/9pGl30LEP0nG22Xg8yLAotSaE563nN0AQacA/+bbQndr3YwlwOsqijEZy92hnwZxBTWT20x7op0Iz9y31zmJYEgkshwjUG0HaBRhfgPFRoZs5XHO9/G0aIGJt+2zXDzsmz4W5af06nZ9Bag3IFl3Bb6v4g9l3fe8/JJ+MBIITS+QihqSr0S3AziFwxJZvrUqsSdFu012gjcMGAtnuf4qwFP6I2T43cWkto4yzXBmVToR16wgJzQ9Dz2UlHImnLNVv1hCpffKlbnryEU/g6+dQVozj5nKAJVEB0PFn2pgVB5TxxmTLPNDnA3DUxBQ81KT3TCpxSjffE4ffy+hledhPeTOu2eT2qpjY90LzNwgEz76qUw+nZxxFMXDMvzZLHY4y7DjD54x28+fXozUiaRjdY1S0zNvxiqm/j/YYFTFqKb/N5IvM75Ry/vV85ymhWBZTsHuQmOsfOdrkBFlQmwlwviK+Va9hGeqsrOojqIKpnJqrjt9t4U2b6Z2F6p6zQl9bfrmFNcIfoRhEV0jYRG0YCoysdR3BmCDmhyOzBOe9YYPilH3V6e/3QThXIL3j43E6REnF7hFiNpRvud+Qn6tpwkNZBWs9IWq/7+9VHYTZnDuct27fxO3md3qX8phnFCL40zATCKu1KOfHGY13khvG7Tl4FzbAx8Z5D+SPAB1EdRPW8RHV0X7QnSCipRKiG8yDV9D+r4MU7V8n13Dn7UgraLeADRUBjeGPr7PnLL8jXbI0eK2JhCyS78X3puFnZnJJCMJb8MF68QlOXOK455nADMhYQNmp3PwFYZY2XMsoY65PxmPfhYckeVAyNwQqH3AL1faRcla0vxY8L03GZHH/fiP/iUt4JHRLRNt+R4g/n4g/nHSkSWTi+BG/LXLy6M7ppKRoKO2zCPr/7yLHECLwqLDSDfLOw98eCpXRQJlvUjpeHwHpMhwjteSXwLUEK6a+sHX6GikILip6+WRmy84n9d8FzOfAb1oPH+Sarz7y6mJCIJJ1nqUpPLE/ssZnuNGC1XR4f9ujkHOkmjmvDLyViGuWqo8BF1lEga1O5JQE/JxrwzGTUYrla5XyI+urNei3/5nOYX7bsSitZLgPlKbMDLzu8vsr4v9NS6K8U1cI7/uRp0s8i++v5us/RD++nLIof40w8iQ== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get all training jobs. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-detailed-information-about-a-specific-provider.api.mdx b/versioned_docs/version-v0.2.23/api/get-detailed-information-about-a-specific-provider.api.mdx new file mode 100644 index 0000000..5a6b2ae --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-detailed-information-about-a-specific-provider.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-detailed-information-about-a-specific-provider +title: "Get detailed information about a specific provider." +description: "Get detailed information about a specific provider." +sidebar_label: "Get detailed information about a specific provider." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuELtsCiZ0GW3TrW7btdgO0SJCPwyINFrQ0tthSpEpSTgzD/33fUJIt2c7WW+wlgH2xPobDmTfvDUUuEke+tMaTT0aL5PTkhP8y8qlTZVDWJKPkTFw6O1MZuXMzscKOv1IaRGpNkMooMxUhJ1E2Jr94kRFeaD9IjhI2IhPYpyxLrVLJPodfPTteJD7NqZB8FeYlYaraNwbCXUkuqDosWaqOkQ8Os8KoH+YNoji7PBdGFoSQlF/FJFRRaioQiE+W0Xd8/LfKfuz11qjvFQnYm6AmCs4m1vUy7rmsne0TKlsIO9kRZMSIvQK9iZrugkdmmWIrqS97QFlDF5NkdLcaYSqt4Wp1P7ZWkzTdR6YqxjGLzag7T6Rzct590ESyvF9u5vYuBl25mIUopUM5Ajm/E7ecpA75c8uwcg5lEnXwwqNgledS9rNbPpXERGpPRxDe90o5AgfvIsH7xNzk1IoNK9Duj5KgguZYu/rcIhs/dEVdDjm2VRBSOJoqj6pQ1uGfSXWVsZ5V8CLtlVGarJ/uAPlxhr/uahhMb86OfBAP0sPzTGqVCRCgkJqjoez/aw51QB07BbdTVGATiD9vbi7baqU2IyZgg+CPBPvBOWYvGx8BPp9bF4SvikK6eVt5ijYPuUpzoeqknZLgCRNfmuZ9nCdSipvkvjPX1jy1tga5ibwqpDl2JDM51iQ6Y3rh8EzKIGeT7pHli4uyZutLcSZurz41yaQIfkyi8mBLsKgsRtOMRGEdGuMWu3hyX1KKbpkKm6ZRLSltxPVf1NFUuK3WCryOAiJMT4DXrnBi4mwRY8AyMRBXtprmeo7yaG0fvLj6451489vJmwGDRo+S23GXX6D6er7fZSauaoqv49mT+Y1yTt/uVk6qVewuGO75IliLwWbeOmavIMJUzQj/ha1gw9iqgg6iOojqmYnq9O16vhsw/TMzvVGW70rrL1tBE0CIHlOijGFjsclAQqtChYG4xLqPgALYI6f4NBUaL127Ur1+aqXy5GYoPuBkKcU1EbWpDD0C74A7amA4SOsgrWckrdfd9eqcmY3IxXXN9vX8jbzOtinfgpENxEUFJpAs4qo0Jiw8xsbtUNbKK6OJrHTYsXv8F8cHUR1E9bxEdbJrtj0kFFXCVJNTz9m0OzbP8ffj/UihKTW8bCMr17i2Wzc+aVlvtOPmmA9B4Ku/o1TsvpTYO+46kDh/v7mH5UKDOJwQT7GuRnAVqrOlzXZXvUROiCW3ME2mFNXKs46S4ezVsPXuh4tOeEsY1StxnUDlGNI8hHI0HOKb4Di32LBmx1rLQh6jHOm3QWqLhOfyBJBVmMeB79tGdHfP7xjAGGJTtU88XlzzeHHdAFkfuBwlPHmNx+zVDohYz77HbNnVWtdzyGVoYfRfjNj4oYYUP9rJZKVFq/Jxhw0vyqEruZqs9V6cG/dEpuRrp9LRFz5409ZFJW47H/PmQ0ObTk7bqD6j0/GJHB9VAUe0DcakIcnPEa6HzmLdx3/SXUOhQI9hWGp8vXGskQSLhjx3dVVW9MH1qMtvFJs5woaLxRhfgrdOL5f8GB+TjsmBy9iOx8wEUCVTnq+zVb95MqcXVw33X4q9pLIzm/akycyZa1JXfIfLbzTfkGo8dMrR17GAcaC1xVmaUhk6Y7cWzJ7wPn64gRT/ARQUq/0= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get detailed information about a specific provider. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-the-artifacts-of-a-training-job.api.mdx b/versioned_docs/version-v0.2.23/api/get-the-artifacts-of-a-training-job.api.mdx new file mode 100644 index 0000000..4d41fcf --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-the-artifacts-of-a-training-job.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-the-artifacts-of-a-training-job +title: "Get the artifacts of a training job." +description: "Get the artifacts of a training job." +sidebar_label: "Get the artifacts of a training job." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETi2Q2GmwRbe5ZdOPzaJFg8Q5LNoioKWxxVYiVZJyYhj+7/uGomzJdjZOdy8Fkktsa77nvRmKi8SSq4x25JKTRXJ8dMT/MnKpVZVXRicnyam4MM6PrFRa6elfZnxqvZrI1LvLqDtIDpLUaE/as7qsqkKlktWH3xzbWCQuzamU/MnPK4JVM/5GqYdiZU1FsNhE8M2Mb+paZR1J5y38QrIf1rVWP2oSKoNXNVFkxcRY4XMSPoYqYCxZIrSc0u+VUdq7jllprZzDqvJUuocDW/v5L6GtQwmBWZKeshvpd9mETslPkgxCh16VtOVohB+dl2UlbnPSGx7ErXQiumBvVJk07ziCCE2RzpbRtnxB4V7LTs4auxXQcdPW/IYbuE/7ztfFMZOttuEH6dvg8UW5jdJV0ucPO3mnChJu7tBiwRqcjKXNbGDceWObbFaJlASbqdvmw7NP4ZMsnotVqaKwkM6ZVIWgbxX8bUZ+8ADKfrJHui7HkGrDvymM60I6Pt408gFSYiYLQNXofgsAOekoVBoCKgtk/mmzaxNdw0i7KuhO+fnDNi9WsrHUQukszBguvsmoEBhAE2ZcSskS1mWWqaZNF50KT2Th6ABD70etuOEnn2PJe5XbTroX7VcIK19wuN3J+DEE9jjnnanSGwcHq7B2sivivxPIWRdj/dqtH60IldVMl3W7ba3dgHuy0UnlPLMzFng9Ru8z9LjkV7O+P6LvKe+uxbOV60qCw5ZiojT5uh0qyHDJAf62a8uNgFMODvM0jDelAwYEBncpC57FlP1/aw5D29duD6L/ORpdiEYaCM8ocLypzkPD7621vHVY+AC1cLmxXri6LKWdtzOXgsxtrjBEVJO0VRJI4YUldXwe/AR0eKmKfT030uy6MBq5ibwupT4EcjI5xlzu6PTCYU9KI2em8oO+OtP4VFxffojJpAh+TKJ2vD4MOgttmhGgjPmvdLNb2bMcm9oH566iFFRMhUnT2lqeJBtxPQbcscNtt1bF66A7lOme4rXHMjGxpgwxnF6cD8Slqad5MUd7isLcOnH57ky8+v3oVSAv3eEk0CCjxRegvvb3h8zEZQPxdTx7Ij8y5/j1buakhQIrRM7nAv7gjYGynreG2SqAMFUzHCZkaWodBks81DyR6olUvxKpjl+v/Y2A9I+M9Mgs16XW36YGJ1AhukuJsnCUBdmwOkWhSuUH4qIgnIewQOdCTrHoRIGHtt1UL+/bVI7sDM1HOZlKONNmDKxa0x3qzYuZYhmeqPVErV+IWi+7++qckY3IxVWD9rX/SK/Tbci3xcgG4lMNJJAsw1YaExaPNuGsnbX0ymgi68LvuPL4F8NPpHoi1a9FqqNd3vagUGAJQ01OHWfTfRsTz85Myf+vjNHPOaF+Au+pqafsv5B173gG4TXWSrzQk2UHi0TjC5Q7L4aKjWGx2vn2HQjMX1+fv2n7E66NjJju8Myu1n3xtkaftlgasbVcIhnElBuIJrDWvm2fJMPZC1lUuRzyS/lhm8sQjocrb5BulnOTUW25yrn31clwiGPCYQ5Vyg6LQpbyEB1Kvw9SUybs1BHqHq5DoPimnU2fv/IzRmuINTbyA+uLK9YXVxGzAct8eQHnTYlmL3ZUjSnuemCXXfp1LYdbOAytmYKRL1ps/GGeUDjHk87ixYDU4YyjLAaVbfCbcb95tFkUiFxjVFr6ohmDfO+Gtm0bH/P7SAG6Wjlto/rIFxHNTQW3AJOEaxJRsyfkeuVYrGf5vvoRLZ7u/LAq8IyjCW1eRJx8TiJO4vXNCin4zlfSGCArtKC1jAhWWizGOApe22K55J8b0DMUMuV4VmarEXNvCs8uI8ifi0eyY2dW32neZ2O41cNPAa5hS4wZjYAo0sBI58txBNzonaYpVb6jtbUre0x7/3YE7v0DE32ivQ== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get the artifacts of a training job. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-the-current-health-status-of-the-service.api.mdx b/versioned_docs/version-v0.2.23/api/get-the-current-health-status-of-the-service.api.mdx new file mode 100644 index 0000000..e1712d4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-the-current-health-status-of-the-service.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-the-current-health-status-of-the-service +title: "Get the current health status of the service." +description: "Get the current health status of the service." +sidebar_label: "Get the current health status of the service." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgU/CuETlvAsdOgRbe5pd1ua7TdBIl7WKQ50NSzxZYitSTlxDD833ceJcVynF24BXoI4FxEi3xf82ZIKqvMU6icDRSy01V2cnzMj5yC8rqK2tnsNPtA0sRCaDtzvpT8EuNcKwztXOiZiAWJQH6hFQkdhKvIp2XSDLNBppyNZCP7lVVlkp2zo2+Bna+yoAoqJY/isiKEc9NvpCIMK8+eom5SC1HGOvTWhegRH+vI1mV2ep2df8SPd947j+dfLopxWRkqEZvy7GbwoKq3tfeYEkVTXeNeuK1qsvV6kMk81001F72EZtIEGgC9f2rt4R/x2wwRKepo6B65MYDLHob/sBW2jy1G/RyGSILTePFYayZYxylQiOJWsp+FNDoXcFFKwz6R2q9rgYbbOfnd4iaTi6405XLgeI/Jbvu2TVP/RFo8EFKEwvkoQl2W0i+77lBac1toVTDfuGivJXrJ0Enbzqc4a/YfpTb7Rm5Wc2jjLGoTRV1Ke+RJ5nJqSPRsttLhSNqiZqv2qPK386qh1DNxJr5cfmqLUUh+SqIOlIvo0FlY04JE6TxtcUROXR0bllSk9Ewr4ZRKjIYGt/P6CQp33boHr8fpTmCPgdftJWLmXZlyOLsYD8Wlq+eFWaI9xrjbIC7/fCte/X78asig0Z1kmfb5Bapv4r2RubhsKL7JZ0/mt8o5ef24cpTRaQeAeeBBdA7Gdtk5Zq8gwlwvCM/S1VjD2OqSDqI6iOqJierk9SbeBEz/zExvlRX60vrb1dAEEKI7RZQzbCw2GUkYXeo4FBeGJBKKYI+cS4jEYNJ3J9XL/zqp+ERD8wEnS4kADxOrtnQHvHFIt+AepHWQ1pOS1sv+eTVmZiNzcdWwfRO/ldfZLuU7MPKhOK/BBJJlOpWmhIPHugjAMNfKK6eZrE3cldj/OT6I6iCqpyWq48ei7SGhpBKmmpwHrmZsGZ24+wH4nhrw1B4fgvwdW0kvS4K42S28YVw4AJbNKalExgI/Rovno8YT3jUHHq9fZbXnzIsYq9PRCEfvUeECsj8yRpbyCFHV96FyZba+YTskpeMyGf7R6f36hueYAYlbLTif2F5csb24anmQ+IH4HLwpdvF8p1uTgmUTtggk+5Tue46FjAIbwULDyVcrHvxBo5TuxmTzymFHCJBgujdoD/H7hhN5+n8B748zqSg0TqWnr5b76nwi/K7zKd/xDSTg5bzL6jM2FBMSdyrgCHUyJhYN+onObuGy2myUP+yo1XukuziqDC5GnF9q/Krlx3XTiZYhaCezgF+vVlNcqb54s17za9zK/LLhWdrXptxrECCZ5thtmRnfacliUIoqZiA+gGoOv7O7r/tsff9uAon8C8EeSEg= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get the current health status of the service. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-the-result-of-a-job.api.mdx b/versioned_docs/version-v0.2.23/api/get-the-result-of-a-job.api.mdx new file mode 100644 index 0000000..1ee8b9f --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-the-result-of-a-job.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-the-result-of-a-job +title: "Get the result of a job." +description: "Get the result of a job." +sidebar_label: "Get the result of a job." +hide_title: true +hide_table_of_contents: true +api: eJztWMFu2zgQ/RVCpxZw7DTYolvfsttst0CLBk56WLRFQUtjiylFqiTlxjD87/uGlGzZcYKk6GG7SC6RKc5w5s17Q1KrzJGvrfHks/EqOzk+5n8F+dypOihrsnF2WZLArEYHYWci4NeVnQ6zQZZbE8gEtpB1rVUu2WJ05dlslfm8pEryU1jWBEd2ekV5gGHtbE0uqLTonAy5aOp7k6Vzcom5KlDlDzmRRaHYSOrzHXfW0PtZNv64sTCN1tl6sPk9tVaTNP0h01RTcv0RH5wy8/5ICqg30Eay/rxeDw5A1ktLzJytInK0kLqJg0O4AkTW0YOyuxvK6O+Ls9//V0hyVnDRkXBmnSCZlwJ5DsVZ+ySUF1JUsmaS5lY3lRFGViSCFQw6RcDlfO5oLgMVX5K3/zog+3i8SwlWBCf5JsFtWinXjIE8nMVMak8DyP5boxwVyKbPmoMIfR5kQQXNgV2kUkzim2w/uNNDpZLCY0RTLFd2a4HpNpU8JJN+K9moqxf+WXJMk7bp3cig7XXxZYpHmr1wOKDfbu+T3xryQXyXXigDO1UIYFBJDSgqBPnTuqYPMjR99iq4nYNp+yn9fXl5LtJs6KIgLkGLxz4l903PnEPwcfKA61haF4Rvqkq6ZbcXUJzzvVSQoUpJOyVNW3vTvo/rxNoHqfR9V06zeWltDXITZVNJc+RIFnIKRvVsdsLhlZRBzia/R5ZP3teJXE/FqfgwedsmkyP4KYnGQ1SQmGPF0YJEBVIhTa5nrJyQU9uEuLivKVcz6NLmeeMcYfm9uB4ky1Thrlob8Pp8jl4Pg7dLY47h9PzNUExsMy/1EuXRGooXk7/+FC9+P34RpUnXsqoTMzp+gerb9f6QhZgkim/juSfzW+WcvDysnFwrqEKUMPf8EKyFsVl2jtkriDBXC8L/yjYmHUdURY+iehTVLyaqk5fb9S7B9HfM9FZZvi+tf2wDTQAhus6JCoaNxYY9TGhVqTAU5zhxIKAA9si5hEg0Xrpup3p+207lyS1QfMDJUiLAw8RqDF0Dbz5IUAvDo7QepfULSet5f796w8xG5OIisX27fiuv05uU78AohuJ9AyaQrOKuNCVsPMYGAIZ3rbwKmkk+Cd+Q2F2OH0X1KKpfS1THh1a7h4SiSphqcu45G75/cei7ob6mhNz2I5PsPjHV0uGKCxH7eO3m+y5fswFqCf59/aJYTIqd1DKUB69zb1518G/MYikbs3fXFHy/66MfXEODm1rs7u98X28DQrQPCwUGHMR8P/OHBQAkAU5pMRVX39geeOlxNlo8k7ou5YizG23y9qNVH7r1CGFgLEW/HrnuVp/OBgnyxnGRyxDq8WiEU8pRaT0KfaS1rOQRCJJ/Hea2yjgYTyi7Csto+KprjR8/8zsWS8yh5dFbthcXbC8uWsnEMmB9Xjyhtnh2AEjuMH5Ha7Kv/r7nUMog0DMXCk4+GbH3h3ZGkXBkitqieXp0q3jEUg590iX5FPxFgzurm8mcfHIqHX0yLAHrYm+46XzK1yGNbuHkvIvqHXqv9lFmNXBEI2NMWhbdoYMdCFbb7eMum5Ytga7DqNY4GfKqsZyrlicfs5YnmM1Mwb8tV/BjvCc0pgsPb+jeMgblZVaww9VqitPoB6fXax7GgdYxHfAYt4Qp1x7kKJTn52LT825N8MmklcNT8SNqPghC99nN8GfR9MFsnOHxKy33uwuL/CcEe5fefyDEtgDx+2CJjRA7PqOaXp7mOdWhZ3bjhLHTOF6fXaKV/AtApGCV +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get the result of a job. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-the-status-of-a-job.api.mdx b/versioned_docs/version-v0.2.23/api/get-the-status-of-a-job.api.mdx new file mode 100644 index 0000000..d36a875 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-the-status-of-a-job.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-the-status-of-a-job +title: "Get the status of a job." +description: "Get the status of a job." +sidebar_label: "Get the status of a job." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETi3g2GmwRbe+pW3azaJFg3wcFmlQ0NLYYkKRKkk5MQz/931DSbZsJ2la9BLAuVhfnHl8896QzDxx5EtrPPlkOE8O9vf5JyOfOlUGZU0yTM5zEj7IUHlhxyLgjqZSV5Jfi2s76ie9JLUmkAk8WJalVml8O7j2HGGe+DSnQvJVmJWEmHZ0TWnAwNLZklxQdX5E+66yznc+OGUm+G4d0oVRPyoSKkNONVbkxNi6CA0RkkUvqfHeF4hMVSTDSyAuSk2BMjxT5jtwTECFx91YKh0fM+qsqq9TaVLSfH21CeZ95RxwCLqjtIqkrLMVIQGTzDLFr6U+6Ux6LLWnHsrwo1IO4QGtYWE5C2QMKmiexr+ItZn/kDN0siuDcUArblXIWyzByfQGDPQBhcH89VClGQj5IG6lRyQUWmUC3BZSg+Ki5uLPFHurRgphJ+S2JvjP+flJO4/UZsQFbgj5mVCOnGNh8Mc9IYXPrQvCV0Uh3Wyp5vjNba7SXKh60k5JFJQ1JU3zPuZZcPwAfTw1c/01p9bWYG4irwpp9hzJTI40ic6YNTicqa3jz3O9+FrWwnopDsXF6edmMtCsGJGoPGUiWFQWo2lKorAO3jFcz9rEcmSrEJP7klIYKhU2TaOsIaN1XL8i5KbCbbWW5HUEHWl6gLy2NYmxs0XEcHhy3BentprkeobyaG1vvTj9+F68+Xv/TZ9JozvJvu7qC1Jf5XsnM3FaS3yF54nKb5xz8PZ+56RacRvIMdzzRbAWg82sDcxRIYSJmhJ+C1vhG+ZWFbQz1c5Uz8xUB29X+c6h9C+s9MZZvmut/2wFT4AhukuJMqaNzSYDCa0KFfriRJMEoAD1yImESTReunalev3gnoTcFMUHnWwlAj0srMrQHfjGwt6Qu7PWzlrPylqvu+vVMSsbyMVZrfZV/sZeh9uSb8nI+uJrBSWQLOKqNCIsPMbGHXPW2iujsax02LbYY4F3ptqZ6nmZav++bE+wUHQJS01OPM/mCPvC7SPgJ2qYWx78ZHsyLqWTBc6ZjofPE4MbDBiB1Bz6u6kPeoqDlDLkWyTwQnf8oaV/OSyWsjKbp3FrOOWK/eAq6m17sVEQ5rUEtDxzPhUKnzkBYrI5818DACZBTm7xaYJQkS+kHiaD6Supy1wOeHaD5bz9YN6lbjEADDyr0S/4yBzbZM115bi6eQjlcDDA9mQvtx4V3tNaFnIPiNObfmqLhFF4nJ6dCrM48EPbEy+v+B27JIJvBPSZx4szHi/OGq9E/pGfk9d0TV/dwyC3Fr9mMtm1fTdyyGUQaJZThSDfjNj4Qx+jeH4gk5UWXdOjTcW9lXJokK72TQaaBbdUN5Yp+TqodPTNsPati01hO/iIz0EabcLJSYvqC5qu9tFfJXhEB2NOGvk8YoA1CuardeOxMY1MAt2FQamxJeSssZzzRiCXSSMQ/pcOe7K38hQ3m+GGw1gn/LjROarKYuA48/kIu88LpxcLfowNrGMV4DIuASMuOTSRKc/X2bLHPTivF6eN/F+K33HvvXNvHkLErDGM4Dtc3tBss5uwqf8A2Mf8/RsQG94XV/gyx8KHFZ5ZrV8epimVoTNsa0ex1ig+HZ2jdfwPfBM1ZQ== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get the status of a job. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-the-status-of-a-training-job.api.mdx b/versioned_docs/version-v0.2.23/api/get-the-status-of-a-training-job.api.mdx new file mode 100644 index 0000000..27d7c88 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-the-status-of-a-training-job.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-the-status-of-a-training-job +title: "Get the status of a training job." +description: "Get the status of a training job." +sidebar_label: "Get the status of a training job." +hide_title: true +hide_table_of_contents: true +api: eJztWNtu2zgQ/RVCTw2QxGmxRbd5y6aXzSJFg1weFmkR0NLYYiuRKkk5MQz/+56hrrbsJmm3DwWSl1gSOTeeMzOcRWTJFUY7ctHhInpxcMD/EnKxVYVXRkeH0ZE4M85fWqm00tN/zPjCS1+683rjfrQbxUZ70p73yqLIVCx57+iLYwGLyMUp5ZJ/+XlBEGnGXyj22FhYU5D1qlL/xYxvylIlvZXOWyjFylWbrrT6VpJQCbSqiSIrJsYKn5LwtZ0CwqLlLgSwsZskki7z6PAaxudFRp4SvFP6BiZNERSHp4lUWXjNDiRl9TuWOqaMf39et+q4tBYGiUqnMJPNFjXCbqTfZBc8yflLlEhPe17lNHD/2cfwS2Y74hLfoS8vxG1KOiiEHnErneisruJg/a/VSXcUl7xGjGkqNWttY/sr9U4QXgdXd4WaiO40oR7naEobk7uRWWaASko2oVAmiap0nK3g0Wj6OAFE2h26zDLIbZ/HxmQUPO0tycdk+29qZ3tvpLVy3n9RW7L8vPyO9ye6ihYHWI5N6YOvpZfVAtE6K1pnhTdNlMJppBR/LYzSvk+Iyhpg31Pu7idpx7mfoWlnSjDMkvwpkGxARqchkKFWwdqoMHHaU4QlNIU7A6ENccOGrZKdnFVyC6TJm4btN5zMHpLKTrrgbEgYeCF9YzwelFsLXSF9er+Sd0hkws0djljwDnbG0ro3EO68sZU3rSM5QWbshoWhz8zG5HqxkM6ZWAWjbxX0rVu+ew/KfvCMWvZV5t9kxvUhXX9eF3KKVWImM0DV6NUjAOSkoxBpLFBJINsPi+1E9AXDbaSsO+Xn98s8a9fWoRZKJ6HecvBNQhmygp4w42KKlsttyW0iM0ecIL+Vig8cdbAK+Urkhk6vWIsC6JXP2Nx+i/AhGPY45b2sspIOdluzNrKrxn/PkOM+xtbKc4f0hlBJyXTpjtuW2u1Hgyx8qpxndtYB7tLoNkGPc77te9p2ZTVXb4nzoBUbeHzRNiKSqyT5sskrcHLJNv6xqeO7BFTZPqTUkOGUDjAQyN25zDgdV53Q/9P1DTq0bVz/+/LyrOmtYpxEoHkVl/vy31trufDw4l3EwqXGok0r81zaeZN2Kay5TRXyiKqctkoCLFyzpK6/Bz0BIB7d4UM1V6tZdWY0fBNpmUu9B/AkcozU3NuzYg5rUho+M5vv1dVLyEfi6vy0dgYdKzoyUbqqIbBMT5oR0IwSoAZdBSt3BcVgYyxMHIemNqY1ux6D7xbU1Wm1wevhOoRpS/CaK4qYWJMHG47OTvbFuSmnaTbH8aDduXXi/N2xePXnwavAX7qT3Af28QWod/r+kok4ryDe2fNA5NfMefF6M3PiTPElIOXWgH94Y7BZzxvBLBVAmKoZ+gmZm1KH3FL3NU+keiLV70SqF687fZdA+gdGes0s16fWv6YEJxAhuouJktDNgmyoniJTufL74gx3Kce911zIKUqcyPDRNpXq5bZK5cjOcPgIJ1MJbW3CwCo13SHeXJupDsMTtZ6o9RtR62W/Xp0wsvmWf1GhvdNf0+toCPkmGMm++FgCCSTzUJXGhMKjTWi3k4ZeCU1kmfkN47/vCH4i1ROpfi9SHWzS9gAKBZYw1OTUsTf9e5h4dmxy/n9hjN4ZjoXfUx3P3m2sP+PZD9dYK3GhJ+vCvFHjATt7F0PFklBV7Xw4A4Hsq6uTN83hhLGREdN1taynOxFvS9od8rOZVi7hBgxKDZZGENVctQ+j0ey5zIpUjvhGvtc4MoLWUXvGVU2ufCktBzf1vjgcjdAd7KXYR8lelslc7mFL/HU/NnnEGh0h3GEQgo1vmpR0/Zm/MUiDofX5nfJ+ccH7xUUN1QBhHltAeRWc2fMN8WJmuxWMyz7r+pLD/A25aqYg5JMWa39IIxTad9JJPRKQOrQ2yiI/2Qq2CZ80ZzQ7kTygDUKlpU+aoccTNxzYUPiYryEZWGrltLHqA48gqhkFxx8JhGNS4+UhSFuJxaLL3w/aXIPE050fFRm+sR3hgBc1PK6jGh71yKYFCJ55DN2OOHCiDATesViM0fhd2Wy55NcVyhkBiXKcGZM2oWw1/tl5Dewd8Rg6bPTnK81XuRdmeHgVIBoKwpgReM3j+hTZm0fhsLbadxTHVPjerkFZXKHW+7eXINt/1Fqt6w== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get the status of a training job. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/get-the-version-of-the-service.api.mdx b/versioned_docs/version-v0.2.23/api/get-the-version-of-the-service.api.mdx new file mode 100644 index 0000000..cdfef04 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/get-the-version-of-the-service.api.mdx @@ -0,0 +1,68 @@ +--- +id: get-the-version-of-the-service +title: "Get the version of the service." +description: "Get the version of the service." +sidebar_label: "Get the version of the service." +hide_title: true +hide_table_of_contents: true +api: eJztWMtuGzcU/RViVi0gS47RII13bpumBhLEsJUChePFFedKw4RDTkmOYkHQv+fceVijyG5VoBsD0kbUzH3x3HP40DoLHCvvIsfsfJ2dnZ7KV85RB1Ml4112nv3JIWKkjJv7UJI8Vdq7RMYZt1CpYBU5LI1mtexMXV3OOIyzUSaG7JJEpaqyRjf+k89RQq+zqAsuSUZpVTGS+dln1gmOVfAVh2TawrrAA8OYArLD8PFi2wqUnw/ryzabUUZ5bsSY7NUgxZxs5BHQ+Ls2gfPs/PYh590oSyZZ3ka/BBJPZh7ChNEw/xgFSAk/PQbzFHaSnmNSXykizpKsyRVClGQlJsr63wCNiVIdB3YGYRcc9qb1x3R6pVprdD0Hhg9w/Fsv3oQg8xfjkSIVCx+SinVZUlj1neHG5mthdKFMO+lgyKUGOnLd+ybPRuKDdfbQzK21pLbeYW6qqEtyJ4Epp5llNfDZKUcyGYc5O33ALH/4ULV0+lFdqI/X77rJaBQ/Y1VHzlXy6Cy8ecmq9IF3OEIzX6eWJRVrMzdaea3rEBjpv6vrv9C363DfrQfwBnRuYHoCvH5dUPPgy6aGi6vLsbr29aKwK7THWv81quvff1Wvfj59NRbQ+J7KqmVGzy9QfZvvF8rVdUvxbT0HMr9Tztnrx5WjrYEqVAH3KIPkPZzdqg8sUUGEhVkyvktfw0awNSUfRXUU1TMT1dnrbb4pmP5emN4pKw6l9ZevoQkgxPeaORfYRGyUWFlTmjRWV5YJBSWwhxbY0pXFy9DvVC+f2qlkR0PzAadIiQGPEKt2fA+8E35xB8NRWkdpPSNpvRzuV5fCbFSublq2b/N38rrYp3wPRj5WH2owgalsdqUZY+NxPgEwvOvklfOcapv2JfZPgY+iOorqeYnq9LFsB0ioUYlQjRZRZnPpBJ0k1e9W+5Zb8Pr75+61T+6hFQUqGXKWQPDHuPCAKFtwowtKBX5Mli8m/bVvlLV7nDisszpIsUVK1flkgt32pPARBZ9YSyWdYKL6y1j7MtvciR/KN2nVOP7WS/z2Tt5J0xs6dXi8E391I/7qpmt9Qwnkf7j0ZssXew2aFqKUuMMZGrJ4GDkVlBS0vzQI8smp7z6QJTfHYXZ55bEIRKiuOSqYAL2Hlga53PZlhQhz0hzboBT4k/wVYH1oOL4ffCbHegvWB1r0Vb3HGmJjQ5cKOEKQgolDhw5q5g4S6+1qeIBrJ+PE92lSWZx3pIamueuOBLct2oPbv7Ranq/XMxyVPga72chjnLbCqmVTs17NpKHoMhywuGAVlfZ/4ZWQXGuuhGe42NSSf2/V3gw5+fbNFNT/Bjr+KPc= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Get the version of the service. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/index-documents-so-they-can-be-used-by-the-rag-system.api.mdx b/versioned_docs/version-v0.2.23/api/index-documents-so-they-can-be-used-by-the-rag-system.api.mdx new file mode 100644 index 0000000..db90452 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/index-documents-so-they-can-be-used-by-the-rag-system.api.mdx @@ -0,0 +1,68 @@ +--- +id: index-documents-so-they-can-be-used-by-the-rag-system +title: "Index documents so they can be used by the RAG system." +description: "Index documents so they can be used by the RAG system." +sidebar_label: "Index documents so they can be used by the RAG system." +hide_title: true +hide_table_of_contents: true +api: eJztWt1TGzcQ/1c099TOgE0ySdPQJxJoyjQExpCHDjCMfLe2FXSSK+kMrsf/e3dXd1j+IDhJH5LO8cJZJ+33b3et9Sxz4MfWePDZ/ix7vrdH/wrwuVPjoKzJ9rPTP7P5TvZi06uLEQgHf1fgg7iTXigzkVoVwjpRSj2wroQi28lyawKYQOfleKxVLul895MnIrPM5yMoJT2F6RiQrO1/gjzgwbGzY3BBRel8kKHyyT6FZIfgcOOyWH9cXJyJuFvktgCSP6igITnrg1NmuHb0yDkUnjfvCCn8yLogfFWW0k2FHYiAGgPvuRupfCRUVNopaYJAhYU09XvmMyf6QSq9Lee4m1hra1A3MapKaXYdyEL2NYjkzJI4xEkZ1NnkW2j50yk/Sf2zOBAfe+9rZXIUvg+i8lCIYNGzeBomIErrANUkf7LnhOzbKjBzP4ZcDVQubJ5XzgGyX5ELBZNFoSK/s8ShA6k97GQUP8phnOxfNh5uvPVgvOsH/0UzPWK8JpbFwNmSZTg4O+6Inq2GIz1F92ht77zo/f5WvPp171WHjAb3shzHyGjiC0N9we+NLEQvhvhCni0jfz4n7V88f70ZOblWiAoxwuOeHoK1eNhMG8JEFQNhqCaA/0tb4R6yrSqhBVULqh8MVM9fL/hdYKSfUKTXyPIptP6yFWICLQT3OUBBZiOwyQBCq1KFjjjTIFGggNEjhxJBovGl69R4e/lYpfLgJuh8NCdBCdA8FFiVgXu0d8BPUJuhhVYLrR8IWi/TenVMkY2Si/MY7Qv+NbwO1kO+MUbREacVRgLIkqtSH7DwGBvQYPiuhlcBA1npsA6xzxFuQdWC6scC1d4mbltAiFFCoSaHnrTBUqd7WG6oabtelfjYFHAvCptXJUID20BLAk+XfNCfshK9g3fCT32AskOYkE6WgEgnHtfRflhH39hiSkr8R1h7ECzZKp2TU9ypUBK/PYkbVTwdkVSkK6NQFaEKPER5xzGgyAINKfZYoqE1cDpAM6xSn+9sePeImPHtunzIBvt+TDKlHMZQrJNfspJqcKjoU6mMDDX6GyDUAguyW0cc6Ds59eIqkrnKGLlM8UmLVk5vyL0M4JoVE6JvIoUMkl9gi0QvInZpG73Z533dGbGb/9bHjuqXFzszejPviA8W260wwt10HhsmXcSmTIMZhlHsxDyF4lPSqu38TmziazG2mNPpIdjY+oG3lcOU9kVphDgnKQPpcyJG9R5xNLnnCHvDglb2s2iQNVnjMjeR1JpGW7Oh6WtcHXlPybmSBZgGHpdihXptkFhRWIMvsQDr2MRVYgrm9zYqfIzhuKbjQa1WGrMMp28DUYD7sIShh4VvghBRiQhiek9HG+5qaH2VQZlNYk8i+HlzhoRlbU2qBammJHdtzukHzO1knsiuxBTOITnbNkW02azNZm02eyKbrWaLNql9e1Kbz1e+iGpAlBUpnaSMrLWSbbfW5rc2v7Xd2veX2Npurc1mbTZru7X/VVLbplt7uo58Vyhb15u4OhhAvJVGfmiHOCCpjdGJ2X19Ol/bqvb80u1jqUq4eSzW1kmdHJ8cLYXRMjEIciU1PJh6sy02dMqm0jot+n1r0ZsmXTJV2QeXriQXpUst+brT59fzNdM+iCYaDTZd1X6JQ9PL4sUVb2KhxNu9g3eH9fYNXm8okcObi3QS7mEd9QbPY4y6PtH9Ot3Vd7I1Td8rHwvXwz09UlV8dZ8cjpfzdHqCFrPupuhvdel9fNhERTzHCb3P822LeKGpS2pSAejFguqE53vwUWVub7z6B26UuQn2Fsw2w7Fk9nOOZ1kPPrtsJaZOOC0qxi0rvVWV2eRWmuks2eYR8dNyYTwSbn5/wxlrQTm4CiJ+RhY/ZmPLv9EZyzDCT93Js25Ah+66OH3pOjncpYWuYqK4M/4gwDOOuLXJRiGM97tdaaa7I6QGxa7WspS7Psj8tpPb2Ap6yNEeYcoHD5tacnlN72jG0ltMY44Wk6VkmnK5MhhZ6wXSpSTZJIuLpDEntssxl9DbGB97PCgcWI6U2tTvSVFxToqK83qgx2Mj9przMXAmzzZkOJp/+qVJoExnkyllbuuwgEwUErkyYuVPCg+MNTAFFwlsTAz/AEQ5zMguDveKGIuYxwcyBx+JSgdXhgZ01vHkcp14n36spQGVoQYjSnWCvY72DHoKn1LygMzElv6rx3NLBkrGcV9PsfY/VfHuWEtFmb1px2PAX0bfpCGPH5ugz3gyTGGPwUKhTQdmM8ozH52ez2kZg9ZN4zyRh9l9iotLyvwjkAVmEYrbW8AtWd0j7F7EdmQidcXlY3XYSJUknjjIcxiHz+69TrB8dnp+gZv79UCzpAH+Pmpzxzrd4TM+WLYw5ztem2VammHFX2WySJP+/gXAdPwv +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Index documents so they can be used by the RAG system. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/inference.tag.mdx b/versioned_docs/version-v0.2.23/api/inference.tag.mdx new file mode 100644 index 0000000..a5b39f6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/inference.tag.mdx @@ -0,0 +1,21 @@ +--- +id: inference +title: "Llama Stack Inference API for generating completions, chat completions, and embeddings." +description: "Llama Stack Inference API for generating completions, chat completions, and embeddings." +custom_edit_url: null +--- + + + +This API provides the raw interface to the underlying models. Two kinds of models are supported: +- LLM models: these models generate "raw" and "chat" (conversational) completions. +- Embedding models: these models generate embeddings to be used for semantic search. + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/insert-chunks-into-a-vector-database.api.mdx b/versioned_docs/version-v0.2.23/api/insert-chunks-into-a-vector-database.api.mdx new file mode 100644 index 0000000..a07cd8e --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/insert-chunks-into-a-vector-database.api.mdx @@ -0,0 +1,68 @@ +--- +id: insert-chunks-into-a-vector-database +title: "Insert chunks into a vector database." +description: "Insert chunks into a vector database." +sidebar_label: "Insert chunks into a vector database." +hide_title: true +hide_table_of_contents: true +api: eJztWv9v27gV/1cI/bINcBxf0dvtsp9ybYcFa5sgTQcMqZHQEm3xSokaSdn1Gf7f994jJVOyHTvtYVgH54dWFsn3/fPeE8lVYoStdGmFTS5WyYvRCP/LhE2NrJzUZXKRXP8jWQ+Sl7uG7nLBjPh3LaxjC26ZLOdcyYxpwwquptoUIksGSapLJ0qH63lVKZlyXH/+q0Uiq8SmuSg4PrllJYCsnvwqUgcLK6MrYZz00lnHXW2jeRLIzoSBiV2x/n53d8P8bJbqTKD8TjolorXWGVnOtpa+MQaEp8kDxpnNtXHM1kXBzZLpKXOgsaA5i1ymOZNeaSN56RgozHgZxonPGuk7LtWxnP1sZK10CbqxvC54eWYEz/hECRat6YiDnGQJOpfpEVr+8ZqeuPoTu2Qfb98GZVIQfiJYbUXGnAbPwmoxF6zQRoCa6E/yHOMTXTtibiuRyqlMmU7T2hgB7HtygWA8y6TndxM5dMqVFYME40caiJOL+8bDjbda441b/3kz7TFeE8tsanRBMlzeXA3Zra5nuVqCe5TSC8tu//aK/fSX0U9DNJr4wovKR0YTXxDqG36/8Izd+hDfyHNk5K/XqP3LFz/vRk6qJKCC5bDc4oPTGhaXy4YwUoVAmMm5gP8LXcMctK0sxAlUJ1B9Z6B68fOG3x1E+juM9IAsG0PrX7oGTICFxJdUiAzNhmDjTjAlC+mG7EYJDgI5iB4+4wASBYNmGPD2475KZYWZg/PBnAglAebBwKpL8QXs7eCXCGY4QesEre8IWj/G9eoKIxskZx98tG/4B3hdbod8Y4xsyK5riATBC6pKEwGFp9QODAZjAV6ZmPJauW2IPUX4BKoTqL4vUI12cTsCQoQSDDU+s6jNP2GSNlfXKH5X3CsQEuIqzevyMwYQGJuzOU1nGXd8AjVuiPHPDS8EoBrpjb2toGb+orMlCvw74cozfsgmDzI7HDtYTmUGbDE1mMbZPeExfKRXEkcjRcno/nfEixvDl8BKOlHYwxJHiutSXE/BOn2x14MdY3vI+dFtxYENtP6QZwo+89EY8l/0JjbNa4m/CllyFxJAY54gMEP9huxSLfjSsk+ezKeEwEsUD2peG7Uj/RKGAysixIIvaAC6JBzw8MVpOHJB885XyG79V/TZn18OVjiyHrL3Gjoul8NsXA89k8p8X6ZEOXO5b8YsRughaeVxAYVs/DCrNIQJPkAIUfcnrK4NZLVnZRLkHGUNoE+5GNTb42h0zxtoDzN8c5F4g2zJ6l9TH4ndqbc1GRq/5ELkHZKzlwuIBiznrEc9GMQXFdLgORYgHZu4ikxB/F55ha8gHLd0vAxqxTFLcPo2EDnxxXUw1L74JgghFY8gonc42mBWQ+urDEpsInsiwafN6SKWwZpYEWJNUe5gzuV7SPloHs+ugMxOIbk6NkWcstkpm52y2YFs1s8Wp6T27Ultve59iyoBKMtiOlEZ2Wr5Tt3aKb+d8tupW/vfS2ynbu2UzU7Z7NSt/V8ltWO6tfX4qTk70dQwaUyGO32D7v6y3FAiwQbe6XaA4aZhlbe6pa3CQjjeg07r/d1229FJlrVScVGcaA38y/hVWRcTYeI30UZip2XdLq7r8bpvindBbACX1ankuFu8kJC+WqP4HLeQSrV77iFZFoBA5Q0JXstqQqIsp4K21/2uNQibZaHSHNhDbTXrX+4JZwCsJUYHGK2AQ3Y1xSMgBiCbS0gK4KeNwKkuqhq1ag5gB2AwDXH6QIuP3kr2trh67c0hMZMhlcYW/U1x9hENhXJOePpZlPBcl6lXRLrlZnf5IQ6cPb7pqBv54/31Xd8ne73BUInHLstH1KPB7T5BD9Wv55kRDBgjrvWdFa7jtpkohaFoRHNmTHv1Mp3WBcIW6HCQNaCYzNmMfYUsLVmXg0W8mScCT7YsC0cBocAeRdjP7WXjAbM15BYqYlCoBmwqlWAVd3mUTzYHFj5CjEAbPOB9Gut4UR1xqHhZMt0gpl0HgZDRsQvExCIXZRROeDUo8CGedZX9V3gqbh0LzCI4LGCVXhxnZz/3qRTeHBHOjK4rCHVFIRWOeZyeCbR6xN1pgID8TZjjBGint4y8JeMAf+2rOELtTn6mFRHDNqc9UDY9ju0mEfoUvI/5H+xm6k6eGfiqtNIfwx3wMTJu57dnp60kIQF2MzPx9OHvTftA91qO5OYLArHCtbaT43qlu1ODf0deDck+s0PdUNONvML5TSJ/XhOVtt1LWyLGPcI7OigPMJC2sRKVi7ajwaNOqha7j3K3im9b+uzmqHTI3nCA2COJ8Ih3D/DbB/nhRaeG79e1UuyxLU4X7BHyh7sHIAzYZbkcP1LKf2xjDme8ldbdT5XmDoeNaPMQVZbHTqXb9AdLXaOcUzmrYUkOOeSt4gVnHxzUwPARaKN8FappKEpA31PfSILkn9OCOKcerAARsmPuiVCqAeihC5Sci04w2iOiMQ6rzil6e8TdaaDRyeRc21wrpS58Q8WZWni85Rp+JpWmq6dYzODX+fyHc8/lTOpzHzMw6u+2Wep36RM9yZ2rLs7Pebk8y4GCyM4UuuHMohuGYDba0gA7gfndkha+br6J7sc4hlcIbjeXDd5sLkl0LwtE31jhSP8+Op3fjEbNWLd7vR+NdzSOPbKdXm7XrE6Lsnnd9BYRue3CP9pZmkf9utkXKapo/aGt2rN/QlQoRntS+mhf/h35z7M44kd0C2iqKfZD2MUI/BBu6xDWQByMGw+F+Q870IGXm2znmg+PLx7FlCkbBpTaTyXr/XFsRHEdNMK0yWIp5wAVaQDHxt/cyXxrDWCe8lRYTxSSzyfMgIo+C5zeJj7Bm9hKgDK4deCleoeW91+QCKKCUyku/WbdsddvOvaI4/pYAiH/YGY+rxRkcWoE/TaaB/W9t3wL64SucSGwwbcIXpyyWiHBj0at1/gaYGmW/kIQ3TyboJ/v8Rs0FzzDiAQQfhZLrGZe5LM7v3Ew56qmT8X+bSH8pvUrLtNUVO7JueMoQ91cf7iDyZNwIwkDHt4avoCX+O9FAg++dlBGpnerRPFyVtOmY+Jp4t9/AJjGu+k= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Insert chunks into a vector database. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/inspect.tag.mdx b/versioned_docs/version-v0.2.23/api/inspect.tag.mdx new file mode 100644 index 0000000..6b9309c --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/inspect.tag.mdx @@ -0,0 +1,19 @@ +--- +id: inspect +title: "Inspect" +description: "Inspect" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/list-all-agents.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-agents.api.mdx new file mode 100644 index 0000000..81b9457 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-agents.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-agents +title: "List all agents." +description: "List all agents." +sidebar_label: "List all agents." +hide_title: true +hide_table_of_contents: true +api: eJztWG1v2zYQ/iuEPm2AY6fBiq755m3dViBFgiTFMKRBcabOFluK1EjKiWH4v/eOpG3FVtME2JcADhDYlo738tzz8EQtC4e+scajL06XxcnxMX+U6KVTTVDWFKfFWFzATBkIWF5m42ExKKQ1AU1ge2garSSw/eiL50XLwssKa+BvYdEgubGTLygDLWycbdAFlUKWELpW4BwsyEgFrH3faihLxYFAXzzwYw2eT4vTm80K02pdrAab3xNrNYLpXjJtPUHXveKDU2bWvZIS6lzImaxuV6vBDlLXFQqtfBB2KmIBYmqdCHRVts4RWKKBGZKzogL/ubYOOxWu89t1+k+F5CG6cSiA/nlh9g9zUBommm5MQzRSXngMHKN1uuM+V9aX8sfLs5goSInek1Vyw5UUXGQ/5FPQHgfEn/9a5bAk6FMvO8XdDoqggubwexTay2QsZmjQKckgJVuxJiclBIFy1NreUdHCq7qhminpGsKQkuQ0f+kjL9fHKSJ15Q68UGYOWpWCyq1BswNK/X8jsw8Q2i5tFbmdEcV2i/37+vpCJGshbRk5kaH6UcfeOcekYuMBQ1FZF4Rv6xrcgonHdMNoc1cpWQmVinYKTEhtNvl+jBNJHIhFT42crDm0toZqE1VbgzlyCGUkYmfNg3Q4kjJUs5FPqPKn8yZR7mcxjgRNxUhKfoKi9USOYKmztBrna0mYRAiODBPbhhjcNyjVlGhlZVKhxJ28nkPx3OF1tzbgdageYfoOeBtCT52tYw7ji/dDcWnbWaUXG4Zf/vm7ePPr8Zshg4b3wGzv8ouovo33G5TiMlF8m88TmZ+Vc/K2XzlSK961SNC0qdCXYC0tNou1Y/ZKRJipOdJnbVsT976gajyI6iCqFyaqk7fbeNfE9A/M9Kws35XWv7YlTRBCeC8RS4aNxUYji4ZmrcJQXNAo57FF7IEZkEg03XTrSfX6e5PKo5tT8wlOlhLN+5KJ1Rq8J7x5IGKG4SCtg7RekLRed+fVe2Y2ZS6uEtu38bO8xvuUX4NRDsV5S0xAqONUmiANHmMDAUb3srxKnEKrQ89J5hHHB1EdRPWyRHXcF+0JEooqYarBzHM1Yzp30Xy73U32jE+yoDVNMDbgM38DDmok+fp40jb0I3YPXPisTIn3fHLntTQz3aL3sBnNuG1xVaw+n/e4Y4wIB9qineHfE99aVCs+mudE4vD9cQrp2M/9T5VlErXOPDc0YUZwVJaMixnGrQBCRT9G81ej5JyupameIIuH8qIKoTkdjej54qiynlp0pDXUcESYyK9DaeuCXXukhqmwiAv/WG9qN7d8j2keU8oMOOP14orXi6tM9ggpxefgqfr5qx5AeG/wD1QCXd12PcdjOO12c0VOPhmx8wf86oHXoSkbqxhaMPHhSDna4VwifslvGBhDNwWJPjkFh58Mk9e6qOp95xM+yGjSuSNgc1YfaNfUPgqkIRxpC2JMMht6+Pug9OV2w++zzc0OeB9GjaZnuM47ldTlm4QnbPTDveTLy+WEnv4+Or1a8eXERG5iqTxvj1tufTehp0ulN8+vuNgTJh3/WraML7Oel8njinkkg7Uit7Fv+QfNoQnTlrhMqNHQIDUxPmnRWEpsuqv2pvED4f317pqk+A0PSD+w +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all agents. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-available-api-routes-with-their-methods-and-implementing-providers.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-available-api-routes-with-their-methods-and-implementing-providers.api.mdx new file mode 100644 index 0000000..9aaab10 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-available-api-routes-with-their-methods-and-implementing-providers.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-available-api-routes-with-their-methods-and-implementing-providers +title: "List all available API routes with their methods and implementing providers." +description: "List all available API routes with their methods and implementing providers." +sidebar_label: "List all available API routes with their methods and implementing providers." +hide_title: true +hide_table_of_contents: true +api: eJztWF1v2zYU/SuEnjbAidNgRde8ZVu3BWjRIEkfhjQYaOraYkuRGkk5MQz/951LSrEcO00K9CWA/WLLvJ+H55Ail4Wn0DgbKBQny+L46Ii/SgrK6yZqZ4uT4qKzEMrZKLXVdia0nTpfS7YQcuLaKKQxQs6lNnJiSHj8ReGwGBXsRDZyWNk0RqvkNP4SOPayCKqiWvKvuGgI2dzkC6kIx8a7hnzUubJSxqGV9F4uYKQj1eFp71TOwCxEjy5gttnpVUXi9PxMkC0bp20UjYxVsRoVNcXKlU8H+Pvq6lxkYwGARKw6KDgISprrkvy/HCM8o5kuyephlvc6ROGmog8oUkAkk1HoujFUA3A86tAnRwhZlpr9pTkfQDOVJtAIJPiv1Z7Q4XWH1X3PW3XfjIqoo+ECL9j0DFTYAuJsmx82QZuigz7KtGUiUgwJ5VEH2wiG5boLNunzg06PQvGAeRv8zJwI34dC4tugVU6U2g29HLZ63qUTKUxf4YY+7qHgplZc2S+7pMeE5KoIMW5lQFtzaXQpQK1aGm6Ryh+nsRBlbIcEhARoRn43zbM1mi0TuzugnlLIO+9ZF2yMqRahcj6K0Na19AuGiRVDyea20qoSOjfttQSlWVLgUR5PeRIhALZ5buZszamNs+hNVG0t7YEnWaaJGfhslMOZtEXPVj2jy58+NpllP4tT8enifdeMQvETEm2gUkSHmYU3zUnUztOOJZWTh4aUnmolnFKt94T0D+r6HlZ3M9zP1j14A6InmB4Br98rxNS7OtUAHh8K6GJWmQWmxxh3G8TFn7+LN78evUl6pTvJYh7yC1Rf5/tNluIiU3xdzzOZ3ynn+O1u5SijeSGs4B7SiugcnO2iD8xRQYSZnhO+a9fapNSoa9qLai+qFyaq47frfFdg+gdmeqesMJTWP66FJoAQ3SmikmFjsUlsm0bXOh6Kc0MSBUWwR86wkwmDQd/vVK8f26kC+TkmH3CylAjwMLFaS3fAO+KJOhj20tpL6wVJ6/VwvzpjZqNycZnZvs7fyet0m/I9GOWh+NiCCSTrtCtNCBuPdRGAYayTV0lT2Zq4LbFvBd6Lai+qlyWqo13ZniGhpBKmmpwF7ubMMjqRq99xNHzs1CVuday4eO27g2f45sETKpFewhJPyHqzvhIoZpRExDcFJ8V4/mqsc0XjnAljeV9kv2XRem6wirE5GY+xQx9ULqDJA2NkLQ8Ajvp6qFxdrG7YDy3ruEiOf/TLwvUNjzFREgX78yn7i0v2F5cdXRKNkJ+TZ0zmr3bcebC6wgbP5JD5w8jpiqGDJXy24sEHUqb0Ct1foGRQM8zK+UydfO7nZXQqVX9vIT195lOzcT7pYjv4hI8CBkrxctZX9QHrjsn3Ag1whIgZE4uJ+vEE2IBtuV5uf3SeblGJdBfHjcHbF3eXaLPsWHad57HjGX51TAMtmE1ssFxO8Ab3yZvViv/GS6BfZN6mZXTCnAGR4IA1D4s7M+wrLVh7SlHDQXHeatPt1MPNZDVk/1/vrqDI/wF+Mx5E +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all available API routes with their methods and implementing providers. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-available-providers.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-available-providers.api.mdx new file mode 100644 index 0000000..21b80ea --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-available-providers.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-available-providers +title: "List all available providers." +description: "List all available providers." +sidebar_label: "List all available providers." +hide_title: true +hide_table_of_contents: true +api: eJztWN9v2zYQ/lcIPW1AYqfBiq55y7puC9AiQeI8DGkwnKmzxZYiVZJyYhj+33dHSo5kK4sL7CWA8xJb5v367vuOIleZQ19Z49FnZ6vs9OSE/+XopVNVUNZkZ9m5+KR8uHJ2oXJ0/roxENKaAMooMxfKzKwrgQ0ETG0dBGgtqtZklB1lvBpNYPdQVVrJuHr81XOMVeZlgSXwp7CskKLa6VeUgQzJS4UuqJRhDqG7CpyDJS1SAUv/sjVUqrPIB0fJ06J+vZMCxfnVhTBQogiF8ptChCorjSWV4bN19B0f/6Pyl73eGvW9RkHrTVAzRc4IMnKPG+89l8nZPqnyCmFnA0lGhNkrYT9T8yF4IM8VrwJ91QPKGrycZWd3GwtTa02uNt+n1moE031k6nIaq9jOuvMk9avzoMlkfb/eru1DTLp2iVUVOGpHIDIN4lYg6FC8tgpr56hNIiUvPDWs9tzKfnXr54qYgfZ4RAr+XiuHxMG7SPA+Mbc5tWHDBrT7oyyooDnXVuQXpOcdsl3silw4nNNsQIre4Z+Rus7jWAheyF4bweT9ckfZDiw8bfqE7gROaPofQyUOjU6Zg/Nsp96hQQdCN9nxgIMFKA1Tjd1Rt15zbr8MTVKWK+eF5OIBPBW2AK1yQYQuQXORmP9/ozIB3FmnyO2cGLVd6F+TyVXLPmlz5JY0UL00gD46x2rkxUcEji+sC8LXZQlu2TIZ45qHQslCqFS0U0C8ZyGDaX6PcSIXCGu9b+S0mkNra6g2UdQlmGOHkMe+dGx66XAkZahmI/eo8qfLKvHsZ3Eubq8/NcVISn6KovbE/mCps2SNCxSldTiwJXJwX6Gk6S+FlTKqX+JWXj/C66bDbbc24HWoHmF6Brx26xczZ8uYA217I3Ft63mhl9Qere2DF9d/fBDvfj15F6WKj8DbS5dfRPWneL9BLq4TxZ/y2ZP5jXJO3w8rR2oVpyWZe/4QrCVjs2wds1ciwlwtkP6XtjZRqEGVeBDVQVSvTFSn75/iTYjpn5npjbJ8V1p/25o0QQjho0TMGTYWGwSkvapUYSSu6D2GEgrEHpjTRiY0/ejanertczuVR7eg5hOcLKW4x1NvaoOPhHegb9jAcJDWQVqvSFpvu/vVBTObMhc3ie1P8Rt5ne9SvgUjH4nLmpiAUMZdaYq08Rgbj3d5K68cZ1DrMHCs/g/HB1EdRPW6RHUyFG0PCUWVMNVg7rmazbGM8x84Fz577CLib64IyBFZ0+fCEkTZHKMugC8IsvHizXhjRo/TLufjTUDtON0ihOpsPKb99riwdLjNj7WGEo6pVPltJG1JZ3m2owJUWEbD31uR393zb9z2SKj2vMn24obtxU3T/HQ5c5Rx8FTf4s3A/Q5rxfdYA10edz2HAkKLiP9ixNYfCRPjCzGavLI0Bnw8jZMX5UjxLhEhndt5KM5Aok9OweEXPgJr6yLLd51P+cVeE+8dzNusPtMU0emAXxGOJEnGhG/V9mhmD4fV0zR80bARccDHMK40ve1w/NjYVUOBu4R01WUaN5p/Wa2m9Kp06/R6zY/pbcstE5vivJpyO+/u04UTX85w87/hkkkuJVbMMzrY1PEiaHtqr7uc/PPjhKj/L5yajXk= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all available providers. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-benchmarks.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-benchmarks.api.mdx new file mode 100644 index 0000000..fce134a --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-benchmarks.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-benchmarks +title: "List all benchmarks." +description: "List all benchmarks." +sidebar_label: "List all benchmarks." +hide_title: true +hide_table_of_contents: true +api: eJztWF1v2zYU/SuEnjYgsdNgRde8pVu3BWjQIEkfhjQIaOraYkORKkk5MQz/951LSbbs2E0K7CWA/WKJ4uX9OueSl/PMU6icDRSyk3l2fHTEfzkF5XUVtbPZSXYqPukQP5BVRSn9fbhsJQbZQaacjWQjC8mqMlpJFhp+Cyw5z4IqqJT8FGcVYS03+kYqQrDyriIfdaM3l7E/S3ovZ5ikI5XheWmdwwQ91uR7c0P02k6yRZo8xRR/B1dd7RXd6fzHE3d8bwY2PxxkZOsyO7nJSpeTwXsoNJkcD1NY6/xdPsIzuxiIjQ/KseDduLYqhfggG3XBxXN0zrR/dxPv6qrxt6xidpsCHhDuNYmcxrI2T0f7SbwuSHT+C3bgQEjzIGdBrGQWSyu3RmBzybNl2IUbiwgFrbSITtSBxNj5NLzUIGgqTZ0gwso2AxFeAIFVNtaNYYiyGe2aoltTrMAR2C5G6UzkdZq0bk5JUW4AcQk4meea50lzsQY9Z+nzGLlfStjaGCy2fB8hjSRtfwhoGQGqvZGlUxvu9wZaSxa3Tzw/b81uw61Dzy2BfCCvi10OjKUJdIAa8L3WnnIGcY9M64xo8b+GkW057AUSeI06Grb/w05knvbwsUQo+9K5gUQlZgmYjfFSWoXas/gprzbN2V7ReFGs+tu2KtgQ6HtNwNmDDEJbmKdzAUNLadguyv+/ghiijHUf9xrLTlJO1s365/r6QjSzhUKUUp1qnHyOvx+9Z8TwZBQDEQrnowh1iZDMOkZTmvNQaFUI3TjttbQxJUja9nvSk4AZpTYv1dzMZtXGWfgmihq5PfQE8IwMqslKZs0c1qRRBRkHz+v65XPVIORXcSq+XH5qnVEwfkRcpnIuC54gTVMC0jzBzYSzpFmOXB2T8lCRAjGUcErV3gM9tGHXzyCyzXCXrWXweiBNYdoRvG7bFmNsDcmG04uzgbh09aRAgRs7Y9xDEJd//SHe/X70bsBBo0dZVg0yOnwB6j2OylxcNhBf2fNC5LfMOX6/nTnKaLBCFBAP/IDdDcJ21i3MqwIIEz0l/JeutqmaR13SnlR7Ur0yUh2/X+m7BtLPGekts0KfWv+6GpxAhOhREeUcNiabjCSMLnUciAucHmBQBHrkRIIkBh/9oOXb2107VSA/RfIRTqYSITwMrNrSI+Id8UZtGPbU2lPrFVHrbX+/OmNkw3Jx1aB9pb+l1+lTyHfByAficw0kkCzTrjQibDzWpZNv3tFr2Vg9aYl/sPCeVHtSvS5SHW3T9gIKJZYw1OQksDerhood2NKfS2NWzV7g+6NKeol2Ec05FrhNrWPhEJpskq5KKhkLvAynb6SpCjnklnC4WoH7z0T9kFrw2rPFRYzVyXCILfewcAFWHxojS3kIb9X9QLkSTTTLwQcdZ0nwz47nN7f8jTOfMNU1iywvrlheXLX5l+3NDStvPJy+2XLpwnQJa8CRfSj3V46FjKJtuMNXKzZ+4CalMzHZvHKoBAHUS+cF7UF632Ah526Zy4QfS0WhWVR6+mo5n84noD9dfMRnewPoeznprDrnnjskzFSII1jJMbFI1+50rrk/X9XBXfNb1kZ6jMPK4HjD2lIa523ub7I293zPhuz3L8sSzDjHPG0+H+Gg9MWbxYKHcdbyswZTqVqNOJM3fHtSoLTwTSHyfk8zhrhSVDHa+L4hXb5s1uxFH5l/f7wG8P8DiJlvoA== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all benchmarks. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-chat-completions.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-chat-completions.api.mdx new file mode 100644 index 0000000..84e1408 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-chat-completions.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-chat-completions +title: "List all chat completions." +description: "List all chat completions." +sidebar_label: "List all chat completions." +hide_title: true +hide_table_of_contents: true +api: eJztXW1v2zgS/iuEv1wLpE5a3GJ3+y2Xtnc+pE2QpFgcmiKgJdpmq7cTqSRGkf9+M6QoUZZsy3Z7dZpZ7G5S8W1mOPPMDDnL/TbIhcrSRAk1eP1t8OroCH+EQgW5zLRMk8HrwTE7lUqfZSI5Hp3MuD5J4ywS2HhRjh0ODgZBmmiRaBzOsyySAcceh18UzvFtoIKZiDn+pueZgFnT8RcRaBiY5Wkmci0tBSHXfi+e53wOnaQWsVo/WoZeH6VzmUyhT5Odq5lgozcsnTANvwXAEAsqjgYPwMkslYFQ21MRC6X4VLQliUuXjWySp7GhIE5DEcEcaSLOJoPXn9bNnqeR6OISNkCB+AeFErlhesKLqPHBJ+V9oTQbC3Ztmq8HTKdMhrCBcjIHsqRiHP5m2OhINrKpd7lFb0nJw8FyufXl0bYu51GLe93gsfrQzSM2L+ERm5hjCzg0M/VSIn+kU6dKVDATD0OJ3Xl07jE34ZESB2B2/y1kLkBfP9mlypU/w0+pcYcHXRZ3Ypc757m+gu7wk8dt4nzCMujKJmnO7HQvUNfBNMdRS/cd8Wro7+GWOyRjmOmmyKPGNjW/du9V1WfJhpl2f8fqSVsGNzJ9P16cMpWJQE5KVGI8CRnwA1augGwWCs1lpHAPVjJdrrFaNXC1UhssqchEEkRFCB+SppbgWFx6/azPzjKrS8/ZqbgVES5hx5rNtSvVLA3ZCU+sPKP07npwAL/M5HSGv0H36wEvdHo92ExPkf2WfhoRA89bqXy9df313izYrfgjXzV+luZPZCQaSm8+gHDMz7XOA3vdLHjBCllta5eXK9sSHncQt3ZrmqJ/BxO9M0RvsaWGy9Zu2ungcyhxw2KZcJ3mSGnJ/fyDodzNEkMMgZQjK01AJo+xpx6jicPkPch7PEXv0RPlyYls7URwvp6bbff5wfidFhh34/ABu5vJYMYCMABndwa20ehT6IjpUCg5yqpbTitsD5fFQW5NP7syCicDmfFEDzcTmskI69ysJbePsMx7u0q35h+3s1JDmkSs62EASIxQuk/0tC57VXMFyWJD9b1P3ZhvOyzNYW3zd8xiKRLZ60ikl7k7rUGvF2cg4CEbTVgMOyJhygWdAWHnxj/ewhaEBzjB3HyCSSEmAMDUIhwa8y6SUOTRHOOC04jHnF1qHnyFjoAjMYdRkUoZSLfEksV1niG34p4jXweGdfgn5lrjhDpNI3DbE5mYvVHPh98FhZo0/EAcujQLrUGiRWqM0JF7CeaTF4HhHKMQswwoCxhGdYz2PRCIQwikNHDfMNLm125LrfosM1dH5t8Uc8euBEcER0vUYmfDrvRxiW2DfABRbgIeRTucdktAvHuvmwTGph0nzh6hIxziqDSohjS4/CYCok1m1uMw35v0YyIhCnFaKYWBzuYKhuV1kXeRGIxpRt/+x267cF2WhiGuQ0VJNWkr8fS4eueP6p1qbqw6Hzy1qQgFPhytPJ8WMbCkNpr12I3CqTJQRwfW1RJGMv++PPvAtssymiZ9BRt9AhQ7oW2BHWtAw63QBopKi5tnBf3gorqEG9qs3p8ZL9/M1rgV1JC95ZCieIaj6ji9m15mFWUbh96SyLGDlZ45BUIdh5gFIxgP6p5V8PS84v9nZBwoxqYDdB+WOEBoXmrkZk98EK8QtjPhbp3+rMcwu1gtMIVdKIp4mlFEpQZu8ip/N0q8TfTeUNgVwTwiS08AyEUGhOI2lxAAfwItwS0qLUYmt2kJmD8BAUI8EsUBDQ1ufu1W46rPUkCoetD5w9M21QVmO/Vi54i/NesPzObfuLU2PVqsidzc1nvcIJZ0ezeIpgBmrd1QXQ1hA9XV0M0o3YxSXU2rbb+vRKmu5hf2GFRXQ96DvAfV1VBdzVOuq3FVMDunsVRgQyEJFdhQgc32BTZeLczOaETFNgRNVGxDxTZUbEPFNlRss//FNrY4Zme3TzU3FEw81mCCam5K3HTlMTujAdXfkNlS/c2jqL9pHxI3VjcH3YlUs5tc8PLln/XKaPvW8QqEp2mWiZBNRSJyrssj9L65D04p/aTHPuiDU0TpFIx1rFZmATgeOuLJyJiPZSRxi7yg4KtI1OKd0bocwUOcLVEFl+28WRjPda/Hipy4Hmo5eK1JEY9FbqOn7MaX037TuxFWGZrq2Tr8a3aaTs+xrRM002xBMeZGLbhVCmthPQ3KPZz1sBMHC7vVwRD0X8XSd2dnDVhualcPyP+kUDzaf1Uk0yHT2SfT2fRwB53kqZPHRrKtvWDT+zuf3REYGo/cDlasq/4+8nNnOe49P+hQWtyK6BznH3rvAfqBelfbolbYFRhO766872QU2VB+YQKs9ICVQVhahD0Dm4+JvGdagsQ1jzPccyWA9lDBYiLpetKQ3UGm4BaB9ex7g70CMxuNaZwPJykUxGSQh5RhmVj2gKJMskLfuIh8FV7Rm4eUN+5N3ki12VRd90Sr66isjmqzyWNQbTZ5D/IeVJu9f06EarP3uTZ71+yVSrIpEqGSbCrJpjcPCY6eEBxRGTaVYVMZNpVhUxk2vXlI9dcURVD9Nb15SKb6K5kq1VzTm4dUV/PUsYHqauhm9InejNKVKNXVkMeguhryHuQ9qK5m/5wI1dXsc10NvXlIIcm+hCRUYPOkC2zozUOCpkcNTVRsQ8U2VGxDxTZUbENvHlLNDQUTVHNDbx6S2f7iZkv1N/+PNw834tCCXfki0UFtT+6BHvc+T+sVnbaiVbT+JfVshL3fu85LY7xFLu36it3BFCgoiQKCmSrdwqlmXN3Eae6ryRjQVPB2uvLXTJgjKPyXMKdZONBbEWz4FjIMjiIfi3mahNayXU42kbnSvWKY0Zsqn8AxPlcmz/MmjfjGc+KQVVOuf0nK9PQxr/rQjXnYvBTSsdGL9zbROHN1422iJ+NaMhVDnpqhznRh2oUjY5EX12DNDklGR7zW1Ax+IUt/Pzpqp6o2/DC2aB6hAhfOIxniYV3MIzzIFH78AON5lkVldnT4pXyEVAUzEfP1jgzSBV10PaW3yOu/rq7Ome1tzmEHNUisU7K3eY7BNnY+wJuEWQqIr4o45vncqZ8wfez1kbRM5xIPWswrc0nZbtbZ4OrWrmx749JRmgBvbFbEPHkBABQas/TGNMixT3thQhVs5jGOzb13fRc2FtVTYrmA0YDNFiZkYg+mTeI+TgGGzFFyme+yNAiKPBew/AJdm5hDucNutyrheYpvxLREeHlDx5GG4/PRkF2kxXQW4SOAUZTeKXbx7oT9/sfR7yZsLI/eff0CVa/X+wcP2YVV8ZqenppfWs6rP7stJ4gkBgBg/AwDY4yCYXAydxObVwM5m8pbAT/jtCijBRkLMioyqkdmVK/+rNeDvJG9R00vLUv5pvWftACbAAmJ+0CIEMWGxoZvGkYylnrIziG8AYI0aA+fcjCSCBpz56l+W+aplMhvYfNBnGhKEAOFqFhFIu5B3hDflcIl0yLTelSm9Zvvr0ao2UA5u7TaXq9fmtdxW+WdMMIhOytAEwSPjVcaC3A8SWpOIUNnXlXMvGhiqyYmoyKjelxGddS1Wg8TMlaCqsanCrkZJRNhJDJonQKZvB9P89s5F2g9nl8IsGVlDgPtSdCAT7RRdYnjwXXm887TtMVUeeFgwW5rkSe4UC36ci9allhfzlV0GEe8no6Y38u4iJl91bzjjENtSov3enpFTH0os5oY+2ByyiYyAimy8Xxr9tM87LMNphuuqBBAdPtBZoDY+WtTOxJcD2zFLc6CtUlvLMwaAVVfN6EXtBrEjioIs5fkNQ6qSh6aZF8ipZZuBLKMT6WpdWpcUuKBJujmLAUqBlNhQJrrGfzh8PblIRh3wiX+htweetxCPxuDWZ02ddKDmdbZ68NDiAZfzFIFS72IsIzqhcIyKnyQ2/znBEqAeUk9NwPfOBf06TO2ISgZ/t3ZiFeGdenfx8L6uLjl9PZlx54hkqsGpnEfZf2ZzcvbZYWYuk7Ywl/gNoRJ10QSZqnEa3BuztPwGC9IcyvPss4J1HHCA6HspDwX13hpGqW5weD25GNMOyNA5RwPUi1V71G97RVyBnIEh1HXAqyGmoYQvP8dyupRpbbhQfthFkEUjiuXpe9WGz5ZGVt9QNm/NMer3BypeloBW4g7jwO+fRtDZP8xjx4e8LM1LdzyUCp0fbXaLyV6U/zr5OOrmHtoC6l9gX3MJcxmlGyKgCuocZi7CzWdELhiTQetO6z5I0BwBcUOmmuKP+MfIDYbIzgAYoC2QSCF94SgV3bQcRCIzJdtK0JtQN4/314BCP4Pzk2Odg== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all chat completions. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-datasets.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-datasets.api.mdx new file mode 100644 index 0000000..4b2239b --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-datasets.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-datasets +title: "List all datasets." +description: "List all datasets." +sidebar_label: "List all datasets." +hide_title: true +hide_table_of_contents: true +api: eJztWdmO2zYU/RVWL2kBL5MgbRr3adqkTYCkGcxMHoqxMaAl2mJCkQpJ2WMY/veeS0q2vMzW9CEB7BdLXO9yzuW91DKxwpVGO+GSwTJ5dnJCf5lwqZWll0Yng+SUvZPOv+KeO+HdeT2+l3SS1GgvtKcpvCyVTDlN6X9yNG+ZuDQXBacnvygFVjLjTyL1mFhaUwrrZdw1w9KtUdxavsAg6UXh7p8tM4ggJ1LY1ljnrdTTZBUGzzDEXkNRU9lUXMvs7oG39MeG3Y5OInRVJIOrpDCZUHh3uRQqw8MM0hp7nY3xnEXrUXdqaOL1pNJpMHAnGQud5gW3n/HsjVH13/XUmqqM+halT0bB4A7mbi2XiQmv1G5b232XkJeZCWv07zCu5nzh2JN6xhM2MZbVLy6YorKlcXdqi37f9ZZLjfZ+IZzjU/gDA2Zc9b9UwtH2Xa7dHI6pm5thTfNoV9azuDGJ63PRyMSkzgK29JRJ7/AK1GUiYxVkhLhRr33gEmRZ7GSw3EROKxsQGhRubQD5jBYfJlDsPrTdhoLGM5WVW16hd4hIfwdm7XiqpXLKNRvDEGMPE0PVCTDAOPt4/rbHXvemPdZlwyT3vnSDfr9YzMXYgS+91BR4o0V6xEI1TMJA5SZx3F4PNQxSN/ttjG1/ed5Zxv/rmtqrYZKsID/PMklScnXWMseEKyc6CCFfKmlFRrgIGkZ94V0vvSKFITY54yI6alfv07XWPud3qg5Tfq2DrJm7LQ+FBqgY/h8RhQ6bZLmPJF0p1RZ8DG4LrttNINUYfGi1rOPOjkD7BliNyEG3I0k65hCHYEipGSm5BtAVWw6ThpPDZMCu8G4hHD0PE7DLDpMOnmowxOY3QinTYXNjVfbDEAKxrVncOZwWnEbfO3W0YqP/BK/gqxa+zvH+IIDtWAIiYKyksYXUHJ3kwRpQi795QYvXexY44sglGLDN5mOk+DYjxS6jjwHjGDC+PmDQ7s0qrRXQWgjPd1LZb9H3e7quRWONBnv50aNM3krIt7PqTuONJr9cp24t4422zXsoo63b1xltEJccRQki1xnjaQqQ0luToTKMoBS0ivlfCG7JnimozqHcc5MLP0bvXfEPVU17ujQdMXYpTCFpGwEC3iDD80N1GXGW9keqzeac0mIoKDNSteAKNilE9v8VaaCor9pxjbLwafDxtlhvLi/PWByNrDsL1KhNct+x9tpaAh4NRpXCXG4sGFgVqI0WTU0gwph5LtOcIhUpbSWiRwABzoPYH/YJ/sXRoB66cxxNWyujoRvLq4LrrhWA5liBDps5W+LQTlJTEEsfoOWPH8qIp5/YKc6qd7Uy9WGGIJoxb+BZzBYzwQoEIKhJ/ozg5WNT+bC5K0UKoqXMpGllLcpIsSPXY/Bbe7jx1tp4LUgHM91iPLuFZJLh9AwpyLmpprlawD0I5XPHzv/8g7349eRFoJ+44UUZkdHgC1Df7Pc7z9h5hPhGngciv2bOs5eHmZMqCVawHNMdPaDixmS9aBamVQGEqZwJ/Bem0iE0eFmII6mOpPrOSPXs5Wa/SyD9PSG9ZpZrU+sfU4ETsJC4SYWgSxba1HIvcDgV0vfYGfIRCOSBHj7F6coUOm1zUv1820mF3HAG58OcRCVBSRV8U2lxA3t7vInaDEdqHan1HVHr5/Z59ZaQTZn0RUT7Zv+aXqf7kG+MkfXYhwpIELwIp9JY4ODRJmTSWUOvdRW6d0l/x8JHUh1J9X2R6uTQbg+gUGAJQY1PHWnTlF/7XxpCqceVapVaADu3HIWosDR5FIrS3MAsyTRUoSX3OV76s6f9dYWIEjZQ3YUivrIkIV22Dfp9HLHd3DhI2VWKF7wL7dLPdPEWrh2dgMzSL8LEVw2vr0bUR54OGGpKSZrPLmg+u6j9zeuvR7R51Gn29MBFIdHDbQGFt6HbXjncqdUFuxtqtvMDF0XIgYXOSgPmu1BsYxVpQXIbfZ+FrzQUByccVXhclFsx1OS/cJnizf7iY8rlFaBu+bSR6j19UXMBI/S5CSwkm+h4K3vQgVvKLzdR7/DomqFe3Ph+qZDKxPvXECmir69qm7ZgRB6ljmW4//xo1WpFzcik7CKiJsSiMfntim5bcgQO+jYJL38WCwJwmoqS8ET3EeGyZjcir9rY++v1JWD9L2smTsM= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all datasets. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-models.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-models.api.mdx new file mode 100644 index 0000000..f1b67fe --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-models.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-models +title: "List all models." +description: "List all models." +sidebar_label: "List all models." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuELm2BxEmDLbrNzbvt7gZI0CBND4s0CGhpbLGlSJWknBiG//u+ISVbcpxtCvQSIL5YH/M9741ILjNHvrbGk8+Ol9nR4SH/FeRzp+qgrMmOs7E4VT6c2YK0v2ilR9lellsTyARWkHWtVS5Z4eCrZ61l5vOSKslXYVET7NjJV8oDFGtna3JBJZ+FDH0p6ZxcQEgFqvyPtVWBENRUkevJ+uCUmUF2mMhno743JDYqYmqdCKXyAlWwjcvx0gitZSWFDzL/lq2ivzlU3E0nc6OKX+YrlCQ6BwNnj/Fx8l7Y6cAEbmQQ9tb4oSs2nWzdt0mmqbLjq6ziBuPel4p0gYs56m3dTTFhx2iSJy6/zy0r3kwbk8c49rIJmbyspPuG62Ctbv9uZs42depYVYfsOkLGAzBrXwVNZaOHT/oZXiK3dbk48D0h9a1cePEiKryIVY2XaznPyVYU5Baw1gCSRaHYvtTnAyhZQx+nqMRawzRaw9j6foKkSJr+I9RuEju3Xdfek4To3oM2ktX1ajvfsVmITXiiy2KDnVQmTpAvbrqebuqodbWziizJaEm1enl6eiZgkxA93JlZev4q23sQHsnwWoG7GVTQLBtHwyXrrVYPlXcqtac9TJvvjXJUsMUedYfAb4PodXGQ77bne/mOtxCBixrX7A2ZSiPGJ2uJGUYbISAm42kk/icm/gip/EwuMcpeYPcnJhuExd92TdgEc8wLH8St9IhlLrUquEOV1Gh+RcWvG7iYbKHpj1YFs7PYhWFY/1xenoskLXIkE6dISvBHo+mDcwxZFgZlhS+tC8I3FabEohtaFGVuS5WXQqWknZImRLSjS+l99BOJEqTSj/WcpNm1tga5ibKppNl3BDxNNImeziAc9qQwo6TJH5Hly491QscrMRafL07bZHIEPyHReMAqWHQW2jQnYM7x0Od+xs4JObFNiM59TTmokAub541zGKi0FdfPoLHtcNetdfF6AI1leqB43ZJATDG4Ywzj85ORuLDNrNQLtEdre+vFxV9/ire/H74dcdHoTlZ1QkaHL0B94+8PWYiLBPFNPI9Efsuco3e7mZNrBVaIEurMcZTcQhmTtDXMVgGEmZoT/ivbQIZrqyp6JtUzqZ4YqY7ebfxdAulnjPSWWb5PrX9tA06gQnSXExVcNiabDCS0qlQYiXOsZhBQAHrkTPLKFy/dqOXbm4e+VJ7cHM1HOZlK8eOJ3jSG7lDvgDtqy/BMrWdqPSFqvel/r04Y2bwG/5TQvvHf0mt8H/JdMYqR+NgACSSr+FWaED48xsa1btHRa71kv7fd/h/Dz6R6JtXTItXhLm+PoFBkCUNNzjxnkzZTHPwwWN5oCal12tB5PpeqpZPYOZJjxeu4iywtSpLN4gFGLUOJm4P564Okw4cakeQ+bv4bx7GVIdTHBwf4uO6XFlvEYj8eC+3HY6FRbits31kP0aqwiIrvO0ZfXfM77nFET7cl3Owuxae207I9QWHnKZ/56x3bdyaGH0BE9kHbtxwPgNrNtP9ixNYPLKS4+iVT1Bac9yBZXBkoB3q71PV4IsADwU1lTj4ZlY6+GO6cdRHS941PeBWvAXInZ11UZ21b+GwLdQT/uCYGDdrdvEHqy8202yXb8jLQXTioNRYw7CW2b9l2+SrVs1qDh3vJj5fLCZY+n51erfgxVk9ukdAS58+EO3bF5zMlhgWfLaK/32jBoM1zqhlH2Kg08Xhnewqv+pj7+8MloPwfVuN5TA== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all models. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-open-ai-responses.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-open-ai-responses.api.mdx new file mode 100644 index 0000000..08d96ae --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-open-ai-responses.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-open-ai-responses +title: "List all OpenAI responses." +description: "List all OpenAI responses." +sidebar_label: "List all OpenAI responses." +hide_title: true +hide_table_of_contents: true +api: eJztXW1z2zYS/isYfUky4zhO5zq95psvSa++JrUncadzE3tUiIQktHwrQdrRZPzfb3cBkKBIiqTs5JwY/dDIIrBYLJ59wQJLfZrlQmVpooSavfg0++7oCP8JhQpymRUyTWYvZsfsjVTFaSaS45N3pvXp4k8RFIezg1mQJoVICuzGsyySAcduz/5U2PfTTAVrEXP8VGwyAdRS6gkdszzNRF5IPXLIC7cVz3O+gUayELEa7h3kghcinPPCaSuBsZXIoXFzPr8l8iMrZCxUweOMXa9Fwoq1YFYS7JorZijObg5mIs/TvC2Wx6f0iUdP2GtswUJRcBkpJpdNciuRiJykwpbQAIgeDE0nDYUzEVXkMlm15qFHxbZMhrAEcrmBZjQ2dmTpksYrc4GzgOkqvhpB9ucy5slTmH7IF5FgNHtmejPddGHHqejDADwMpRbImTOZJY+UOACU/V3KHKb+4oOeXc3QJUhDFhFy1MQYzQ9Zl+Ew17Cof5eVIKTI2RLYLtZSVQtBUoCxo2Fqb7GZS6xUIiSK9WIiObN+HfRALRQ8qNSLhljyMmp96Y6r1UovXz36AePRNd8odlH1vJjR8GWRlcUutUkTcboEoQ8qUK3FrS5mTjcHtx+mEB87pdWUwjkiGFoywxUiGdEmE5ithSLOX5PpFz51mNOgrvi3vnaHfmlG7F+BujOuwRTcmwGJ2V7QnyD5t3qKhpnzLkbPXfkgMBvCASOUME24tkPQKuZgs52F7DPGZMl2gaGWcZReuxRbz9dytd7ZgJdFOru5dBeIvtqe8RtxBToJUNDc6UnHaJSA+QDmDWQPWMATthCwTsDXxewAPiAD+AmaXxBlrTvjsEMDdICn/n4f9FBvzQZ9nJf5CKPkOJzf3r2plIJEYPV3EiDNMu8ByZPO6Z+4rEwFJQJAIrVYJrzQDtegcvMrj5EtYhNsOMQZKB9o4GjyYIzgbc+9sD0NBRoOC70l8pboC1si5Hz0+DjPfXct3RZpBCJ03Nc2Lu73KP4kSQuKVdVdxIhDXC1hbzMPpB6xwVf7SWN/WbG5A64NEhqw9NV+e4PUbtGWIhdJgLE9ELNEE3I3Q1TRKe0gJJNQfByxGz1LFekIow6WoJ0pu5awh9H70700y/hMKytngpbFXm2r1+Un6POyb/nwYc1vjTlSOysY3DCqTAQg/oBERCpY6Z6Z2pjYdAiFYMC6Qdh6MBmDLgUNQZGE87EL/ToJWWYXe3uZVcYRAc11PoDZ8bwYPcJ7bD15DLP4g4ERNuvA+7VY4DqmZR4Q7kc5EMdr9JLaA+X1cjRFZyepuRsB+F6wI+NjsA4WWOTAd2NO6i7wjSvHZSLyeb+97Wtzc+A86rKcA5DeZXR32s5BKO+z2o2pNBe/09y5HIzBgCXfsH63X0DiLePFuu0i6dsBId+lBGsxTXIFZ8TmXhu2Jh6HIyQfcviQ47OHHM3Q4Nag9BGIj0DucwTSGx7cGvk+Nvk/xyZOGHE3vvUbiFP60znHTn7kZu90ajPP0svaKeVmmskj/R3lWG9uLoGBPNWGcDjZqTaqEPHOdGaIaUqcxs5WpRpowJWSAE801ZcjclM2Re6Cxkmb96AFBihK1fFo2sGydilGjkPZxLcOp83caa6jhFCxIiUXccVzmZaKmS7koZX1ZJaeYsdnJ4fsfC02jOcCvHbESsArxH6JOZJfiICDwNvd2EpewSf4PqZ+2EBhtHdBk7iYAQtRKSAUAB8OD4G8VCxJGQQGxKTQnaB3hi4YWAtAUmAjmArAyAD7o3aftzxqL9I0YgFMYNeatuRdgryTgun21iuj21KC58GaIYOVhRvCH/Sb635zYsTFYdezRoRh2d8Rgm3RmHoQQ1bMSGYIoA2T8btYvKdhkcmXXdz/XkusWgemU8LV/Q1cqa3E9xfHBTTKGz6oN29e24etAxgwSIgTM1tDEEIeEZTm0s7twIcOZy/0kafqg1/nw6n42yaitwGwmDDGGJn2LDAvYCoLkF5n2254d/ippKQ1rv5ewIwEb6RNkjJedLqc1kFKG5bkgXrPon4Rm6dkJlk9GwbuKw0k3uWijW21vHeznSegmGDPXorSa7H/zn43TRWkuUvQiHOb4DsRiSsOmxpGHWpVNKjWBNnjhSiuhUjYEXmW508I5ONOyjtOyRvEpxlGB4B9MTLN3LA3zmhiFNi0mu+MorR2ri7neGesKPMEMLPY9BuEw7ZlctC4D8U9fIm1pvt5lbaAupMtn8uvIK09NVAjGoYuk4B2wNa/7KFxlsRCoMIhIe1FeL4qYxhxhCP5z/vTX5l+tq28Ffma3BhXYnp1+JHWk8lOxKVgjv9HLIKD7uMKpF1rslcc6FDv8cruShtvP1lpLOIMTtw1nqY5hpl+vbHc3rdg7O3Ls2ptBndzQdbGX+PLqdCznTXq7kbB7IyaCjbdDiCdWBTrNGxZAtglX4l8HvHFmPvDb7BZ61Y2ktd02Bq8bWS/t4tRXTQfrSevG5ez5bIpDXPbvOui8DBpDXYbJyzzNNYevgzwhtOyjBwcTXZaNnHiqJ518q6cx2ki8NGpgPoSN6VaIFIBtoo0AHw+hvZP7ps+RrijQUOhpu03UJuw75z6trR069E+ulqT0Bp7x4oAkrySYVW3QKzivOnD3rsZfbmtv/bkHu9ntJkjzhksZm3kUCSPlLnIlvEc9KUQ+Z6WDom1otcJ1uFV/QBJXq95UTv+MBVqolFw18vMx1F+wAtu+RGyCv/XDWYtLWLJDbb71Ksjerd5BX4FlpOqX3QfDdI6dq/xu7/ta+iRBfxog1eJo231tqbbNHGO49yeJJl4njiT2/dWdrvUyFed+JvfvurE3/W+d3e9fdXJw7A9vurEW6L7bYl81Ym/AjpxE+evgPqqk6GF9nc+/Z1PX3Xy9d/s9Lc5fdWJDzm+pZDDV534COQBRSC+6uSbjU181YmvOvFVJ77qpLfqpFV4MXyBw9ehPMg6lHaNxBeFii9N8aUpvjTFl6b40pRvvzRlq4xi+LKaL1XpdS2+VOWBlqrUtSCfO07zVSu+asVXrUxVTacA5EsoqC9j8WUsrt3zZSy+jOU+lLH0Z+M1R/0yNL/3AkIzLJDqssf2TudBHTPBZ1EEh7RPRWCDE47I5FDk4Kq7VT/Hzhkxb2XS1gKzrs4Q9n6vzSdhytWORePm4gpzxnPrCkbtWByNOHlldcuSqg+vYSyOogeZqsq03y6b1fEjSXqbHxO+yzH5BHdjy+NMByg1gb7f7rGphI7Ugb6UW5iYbSlXZV4fd7psWz3oteCa0vhjseEjGHPNtr8B/vKWtUGD7SoD3tK/EvBPF8n1xX+8Se5Qri6UO1QuZngmYJ2k3gbR9Mk1jzfwWAyRNO6cNO9Ks9Mk2tSr6nB1qLNQX6G7xDm7LrMxb7VOyygkKIIEQMiH7ARV8WwDQX/C3r/65UCHYBL1qhD47I9sE3JYiOAPRr95NSS1fT3oMQubPnSfFYNRGr+i5drHnnFr08h0dxhAoHwCfc7lCBOkBWzleI7VYC5GiAN7gc68GsmLjxyJ7eZ58mlyS7+cqTgmZ8vc6ItEFG8a72OaGfpm0zjETbfnOzf39Ys0m2eTrOyvZRAJsOHKWtsqjOu1tXmZBNvXPYbBdV51w0UGL7zaMPqRQRjEnJG6P+5Gh8qTykigg7uX6UrIo6iFKqoaoruoafCnmw/zdNMfZfqjTH+U6Y8y/VGmP8q8y6NMf3Tpjy790eWXUpodJ1NTATc3hHpxZxt87jus1XJVDO1eJKpJ316c03oyzdwG/Qh3BtJHABDIzBphknoLcAXm4FcC2mUQVKAl58FfdrenUxljFs+/WSh9OG/3GFom/z4P/z4P/2Yhb3v8m4W8JXp4lsi/WciX+U/cqvsyf/9moaGF9nX9vq7/DvDtq/ftavs3C91Wgv7NQj7k8CGHf7OQj0AeUgTi3yz0zcYm/s1C/s1C/s1C/s1CfW8W2p1Z7b/SqLOnupSHDpwje6NY1vUuExcrF3j9bc5xvUi36bAa/q3sVXW23lUd5N4s0tqorxv3KyCR/R2CWz3Z7blu54M1G+RHk9De09OCCEy1qUx00ljuLCvboqiqG38yb9JDEmuu5nHzLl3f7f7qSj/BARGHHaurXXWh10JsUoKNxEsCK7PpyFUxaitz0nG9EPoSGrQKOGQjvjdV7NpN1ABih9HByreGxam+aKywXtH+8Bp7TT1ZC3mBtYvVujmircVRzcFBJ8KjC6Ht7Rpf4Q4bIBgZRHVD1QAr4VdypWPCWBQc+aNDDZjUP46O9KHa9omnubQP1gStGpgaGeIpVcwjBLgITURjbohQSYGuM3iGdR74XX8VUTPqaBn6vq3Ez+fnZ/amVQCWYcJGQdfLU2NYXCwHAouoyjjm+caijarw2fVaBmu0pDjpXHJzWAS7FP2cxiHFtseQY0bWrXHoKE1W+B6AMubJUzB4IeljRwWQfikAhU3oa4MRs2wUFeGuQE/GHDtSYQlY6FxAbwgEtG1wDBbjC7Cv2tHYXXIaBHTFLRBbfE1RiNosmx2PEZ4DfRJTj/Dq8zj7MgLyqu/ScrWONrA8UZReK/bup5fsh38e/UB2V3zEyhrh4gugXo/3Lx6Cu9V1KfVijkO+0ZzvfuzWnCCSeMQI6s8UHbimKXRONpaw0lWg6N7h3zgtzZUCGQuvVF6pvjKl+u7HerxzQPpbRLrRLOWq1n/TEnTiCguvAyEwfKJMC3gxcGKxLA7ZGcQ0GAkDevgK49QIHubWU33f56nM+yNAnKhKEPiECKwyER9B3ugjhRGDVy2vWl+Ran3v+quTxOTu3mu01+Mb9TpuQ94KIzxkpyUgQfCYvNIC62iSlGLd0KpXFShvq9guwl6pvFJ9XUp11DXaCBUiLUGo8ZXC2RyvqBLjsnOHj+madr2ETlqYd/VgIk8nn2d8WRDOJfbHKqlN+5wDplu/VoS2xZU89GKWeYIj1AI3K9DSvzqLVjFA7neYAV3A56Yv1NTBnYxvNbrN8QyMTs1wvKWMsEy+5mGx2XvmaR6OET01o/ckoMVwRwblAluKX+uCuTqFxR4/4ip4hDuHR0jx0ZNJXOIBQRlTvaEKDFONRJbhfOuwDfnT3KK9yqo0gVu2gyk//bI76LESOpuGxwOzZ1fPn4EOJ1zip6rPzL6jSMOWDtNm66LIXjx7BtHe03WqYIynUcRj/hQULfjrMEhjSi0qAeojiw11fGVdzIdLfIZGhyZusx/Yn73H/uy9MT32sAgH11O8et6xRPrFHa7N4q4VdSlTrtK8xkldJGzrP3ALgrZjIgmzVGKxjcmvyhz8jU0D43Eagjlf8kCYBCjPxQWewURpTja2TXyB28oIrG6OGWPNFb2sTh/CZSBHcAgoE4PPndZk630jlSPe3cvADJOLz7IIomznfFTD4IOWsQYCyv45AdeCAVYOFxzbffq0gID9tzy6ucGvtQLhSodSoUcLu1/K5PI62rJ18v2X2DgGlPLh8DedXUxjYdC27Rjems9bDD9k3HaMbs3nLUa/O/O2g09rams+L/EPCK4WqP0f8CBpDZEQvn8EEKQ7HQeByFzZtkLMhjH79+tzMG//AycHQK4= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all OpenAI responses. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-prompts.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-prompts.api.mdx new file mode 100644 index 0000000..37c6c37 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-prompts.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-prompts +title: "List all prompts." +description: "List all prompts." +sidebar_label: "List all prompts." +hide_title: true +hide_table_of_contents: true +api: eJztWG2P0zgQ/iujfAGkfQOBgNXppIXj7pBAu9oXpNOCkJtMG4NjB9vpblX1v98zTtJmu+Uo0n1B6n7ZNBnP6/PM2J5nnkPtbOCQHc+zJ0dH8q/gkHtdR+1sdpyd0Dsd4pl3VR3DeSdOubNRaavthJQxVLefD7K9TL6wjaJI1bXRuRJFh1+CaJtnIS+5UvIUZzVDvxt94TxiIXTU7KNufSlUHEop79UMQjpyFX68uvVnIBeih6+QuxvcZckUZgFKuxAo8m2kGx1Lmiqv1cgw1UblXDpTsA8H9KF7HUh5JmfNjEJT185HLuimZEtNkLREqO7TFejk7O1BttjLptCh20x0nmmka8L+nmsfWkl62AlQiArxScIjPd4jbXPPFTINsxALasqPxEIbx2dd/Dj6K6u/NUy6gBY91rAxdr5SUVSqQA9qUfTb0xf7hZ7ouF+qUP7+IEXR52CLCnW2F+vGBVXkxn3el9m2qkK+Yokoc2VpxMgn/NE2ZXRZpQpViSxqdfhc8Fg1ZljvkXOGlU0hd9/GygRe9+JVKwf1RUIq0osawpKHOR2oqxfpkMx3ypavka9WrgPcAv6ootCiXZmzASY7656/NdozinO9xMKwZsPc3gnt014WdTQSW8vFe9U86ZMDTrvG54yHGs9SXEENAORgmU5rtidv6TVkEXAC+N2kSq7fGVUpuogq/wrc/lRYibkDdze0j3u+L/tK5Qo2FB0ZgceyrSySC0839SehsJhnyN8AtdpOldHghKdKGQE0F/9fWwIJYxO2oO/fl5dn1EqjWRYJqV1GfkTLN94LrER4T8pWoregx1SV8jNhjACRk8xNqfNSsClBAzY2JkQCz+33ZCcxD83abGu5lRbTxlnpPGVTKbvvWRWJoYM1d9xJZLSI2eZbRPnwtG7h9IhO6Or8XRfMkPOAgWes5qkAwwsw2wYlltXINTEZDzXnaF85uTxvvGeYX/PrZ+DbVbiv1jJ5A0SnNH0nef1ApTGwm3yQ1k/nrpmUGBVjZ4y7CXT+52t6/uLoeRoKfKvAPB7iC1Bf2XulCoySBPGVP1siv2POk5ebmZMbDVYQWjtJo0DKHRbbWa9YtAIIEz3FYFOVa2xq2lFXvCPVjlS/GKmevFzZuwTS3wvSO2aFIbX+cQ04gQzxbc5cSNqEbDIcja50PKAzbBvgUAR61AQ7YZLJ6ftJ9ex7kyqwx9gnpFOoxDKPUZvG8i3yLfsu7tKwo9aOWr8QtZ4N59VbQTY8p4sW7Sv7Hb1O7kO+T0ZxQKcNkMCqSlNpxBg81qXzSdHTa7DfX9sD/4fiHal2pPq1SHW0ydoWFEosEaipSZBoutOXeL/hDL52f1MrjxM4CCxLsQLPpUNSsgknJqhY4sfh9PFhtwgv26kmC+ZZ48W9Msb6+PAQ83W/dAEu7hs5Ue6HdKLMXZUtPsk6OKzjLC38oyf19Sf5JmVOAOqPkasTKV10xU4gyAa3Ktn08YY7HuFGuIMSNcTtUHO6d0BgUw0lHy2t/YGInDbAbIvagfYBPEubA+3BcN8WvpDztvQEP1Z5f5mhPH+UOzOTDuHR3Vc+ko28Ac69mvRevZfzcEgAqZFHUFByInck3ynfndjnq463Ubgjp9x5HeLor63YSQWcd4W+bjNarxAk5ZT38/kIG6ArbxYLeY09lJ+1iOlvMFIlsQAtA71RSvyVZwLdPOd0f4HjSpOujtZ78WKIu7/eXALQ/wLZ60Bm +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all prompts. + + + + + + + + + + + + +'"},"variables":{"type":"array","items":{"type":"string"},"description":"List of prompt variable names that can be used in the prompt template"},"is_default":{"type":"boolean","default":false,"description":"Boolean indicating whether this version is the default version for this prompt"}},"additionalProperties":false,"required":["version","prompt_id","variables","is_default"],"title":"Prompt","description":"A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."}}},"additionalProperties":false,"required":["data"],"title":"ListPromptsResponse","description":"Response model to list prompts."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-scoring-functions.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-scoring-functions.api.mdx new file mode 100644 index 0000000..df1cfc6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-scoring-functions.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-scoring-functions +title: "List all scoring functions." +description: "List all scoring functions." +sidebar_label: "List all scoring functions." +hide_title: true +hide_table_of_contents: true +api: eJztW99T2zgQ/lc0frqbgUA717krb5S2d3TawgB9uClMRrE3iVrbciUZyDD877cr2Y7sOODU4aZAeCGRvb++/VZaWc5NoEBnMtWgg72b4OXuLv2LQIdKZEbINNgL9tlHoc1pKJVIJ+/zNKRxfVLIDYKtIJSpgdSQKM+yWIScbtn5pkn+JtDhFBJOn8wsA9QoR98gNCiYKZmBMsJZj7jx7+JK8RneJAwk+n5pEaELYixAefdqQ04Ht/bmS7xFDTFgmasQhiK6+8Yl191A88JWAGmeBHtfg0RGEON3PRUQR/jhEr2VahiN8DOFqIGc1w7P4bgAFIdGkIbThKvv+NlIGRf/hhMl88zFm2QmuLCAa0PGF5VEMOZ5vPSin9izKbASDUbhbDEeX/GZZguitw3ZFlwSMLyRwSpTPIoECfL4uJYzmcLRGEGrJNI8jlFZ9X2E8QNP/SGEeYQ59kYqHxrU8QYKT24vbtFTBSZX6bBM5IITSwi2LPFVMsoBLwXzIR/4t4K+JSLlSAyL/IDtO+TPC5nzICBX24Eb81gDxfEjFwoiYp11CplhhInJu1Or5YxGm9aPueKYLHCW2RhdcDbZJY9z0IMW4FbFokiSj4U31BkLJ9MXi89WS1cs0CYoEa4RjJLEPhr+WGc4CqG+eLxxaroCUlhdIyDlzD6HYz7SGQwr0heKfVLSFQhrcY0wVGJzHLyhzkA4mb5IHFktXaFwNteIhe0VfCSqgc44kERfFD6gjq4YfDg9+rxGBPK0uYDPRzpjYEX6gvCFlHRFwVpcIwzhlJthiI1ODGR0KNIsr9fI8js6w9Sqoi9sB6j0oNJ5SCq7gkj+sLk/zPqzFjTvBLIfhuuG7yehewDU+AS3EkPbJC6i1nqx+6rVkO69gJG+M1S3EmjWC0ZeVKChzsh3nPApUJt9RnlUZJz2BPd4hBft7Bxy9+7NnlPrXDa6vUF5Uj101fD2huXJddNF89sbmCfSVZfh98bjKbXXQfks7bl32UVT3BuJJ9JtL2mGe8Oz6cL9fnKxU+4P8LNszBeb5/6r3jPs1cmLUqmVsLro7IK+tD9ZvwfeEs44ToZcD7/l0QRqUC5caJ4kWIfluDw7YOXZAcvKmHR1vtBQ1sxteZTSuE3BBK6HqE7bHnjEtQhrjXtxRJVaTHQJihUfumOZFiLVIzmsTpEoFoNxffz4iVlhZiTLNTCuGWdWqdsAOKvF0VGSIWMAmc9NK23r1n47yhyTfmcHuTYyYU4FK1VYC+SFs+eCqGIi0zC0sPjZXHpyNj+sqbtx4jRQgHBtFMfWg2zyVF8hDmN0iSEJQaFDESsPDF05T9C6Pe2rzopWcGSeaX6J2m2Sr0BMpmhnOB9KIBK2uQ/RgQmiHfIY1408tUdLYZijx7N6oVVuleeVrdVGg5ry7MVR8Vb72aWosQ7cmts4Hm2VRSjpKHRGHwhLmyprC3g4ZUpe/cRcUudyOwuWJcWDBxm9rz+QbKNilk9HDgwU3OZ623Fxoc5xChmLCSaDvq3yGKacexr1PZ97Fi70mXsayh527iEBOkhdU41uKvSXr9BmxjvUo03tsaXWigVpjTDHygcoSMfy2oOWaqRPCZZaHrb2NtT/n6l/L9HfUOZWpLjN9v3c/qnH1zWedd6JbFrlTau8WYh/+dnoUbXK9YJdeSradM6bzvnRF+yv3DkXtF25MDcd9KYEHkkH7T9UblpYKRbvRfz62/RbVQtevipefxe7jZALwe0vRlK9w07RAh2OYUx4g2tFZW6y3KZ2pSisf34ncMdvIEg16v6j7dcT7iX7Hzlow66wHxYpOigihq4mPEaPkf7B2n5CoQ03uV+KAtVOWtb+f87Ojpm7G5kQ2XIvQr2vE3+nFHXadDPOR0xPpTJM50nC1azcCYC952oqsBiEC1phkRt3opEW100xz0TIB9Fhx+Esu7vJdCxTjI1N84Sn2wqQU6MYmCdTc4csCZyWeRqutt/YZ19OPhbBhOj8CGhzE1HtI32VgEvabijAMCmfbq7gIySemxsyCLEcQiZpLlOA5ht+rcLLIsNltirwPKpamJaAV3Y4rvMhH/aPDwfsROaTKU5nYxnH8kqzk/cH7M+/dv+0EyJcczrK8/mFVPcnpYidOIrP/enI/KJyXr5ur5wwFnT6NUVxbY/BpEThdFYqJq1IhIm4BPyf0KJhsRUJbIpqU1SPrKhevp7bO0OmfyKmF5Wl/dL6V+ZYE4gQXIcAEcFGxUZPQWKRCDNgxzFwdMgge/iEY5HQIxI1KOrt1bKVCvtD7MsYwkmlBAgPEStP4Rrxph0SFDBsSmtTWo+otF7569UhMRs9Z6eO7XP7RXntL1K+BCMasKMcmQA8savSCHDhSaXtd6OyvKrd3sJPae9QvCmqTVE9rqLabbPWoYRslRDV+ERTNM1tFYVRd5s2X4zH8cLeTw+C4q0nu99FbRd2ezmViFMwsb+0zriZ4pedyxc7hfj2fG+9FbglT9t3pXJFvk+NyfZ2dnDx3Z5Kjf5vxzFP+DbGHX4fhDKxx00aMBphZlbwbVnxXy/oGnHAsqvcPJI8OyV5dlowwTIkoJ+HK+2ivHzR8lyICkfXKMR9UvuaDb2/WGy49XnKGn9YpWC7Y0ijTOKcoLEIbecg6G095VgREbw0YagxD+mgg5RyBecpZVYqS/lF5SPq8mNwD3UKrz7R/ts9U8kQx8S9/Z+607i7U9r4sXk1Md4jVlSzgWuzk8XY9tg3hpWdVRwNvjqUF4mAWaNk0x03NyPsnb6o+PaWhrH9UjPHLDuBjSilmGcUwNmGHtcjAb7DjFgfhpAR5+wbuvSErDmN3/r8/PvdGdbCf6vLROg= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all scoring functions. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-session-s-of-a-given-agent.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-session-s-of-a-given-agent.api.mdx new file mode 100644 index 0000000..d2518ef --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-session-s-of-a-given-agent.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-session-s-of-a-given-agent +title: "List all session(s) of a given agent." +description: "List all session(s) of a given agent." +sidebar_label: "List all session(s) of a given agent." +hide_title: true +hide_table_of_contents: true +api: eJztWFFv2zYQ/iuEnhLAsdNgRVe/eWu3BUiRIEkxDKlRnKWzxZYiNZJyYxj+77sjJVu2lcYB9hLAAQLbEnn33d338UguE4uuNNqhS4bL5OL8nD8ydKmVpZdGJ8NkJG5gJjV4zG7rwf2kl6RGe9Sex0NZKpkCjx98czxpmbg0xwL4m1+USGbM5BumniaW1pRovYwuM/DtUWAtLGiQ9Fi4rtmQZZIdgbrZsmM0Xk+T4cN6hq6USla99e+JMQpBtx/pqpigbT9x3ko9az+JgFoPaiSr8WrV28nUfY5CSeeFmYoQgJgaKzw9TStrKVmihBmSsSQH97UwFlsRNvh2jf6dI1kIZiwKoH+eWNuHOUgFE0Uvpj4Mkk449Oyjsqplvo6sC/Ln26sAFNIUnaNR0QxHknCQ3SmfgnLYI/78W0mLGaU+1rIV3LiXeOkVu9+j0B6SkZihRitTTlIcKxpyEiDwhFEp84OCFk4WJcVMoAvwfQLJMH/pIi/HxxCRqvIDnJB6DkpmgsItQLEBgv6/kdl58FWbtpLMzohiu8H+dX9/I+JokZoscKJO1XMV+2gtk4oH9zgVubFeuKoowC6YeEw3DGN+5DLNhYxBWwnaxzLr+n3wE0jsiUWHeo6j2bUymmITeVWAPrMIWSBia84WHPYkNcWs0wOiPLkuI+VOxSgQNAaTEvgJisoRObyhytJsnDeS0JEQ7BkmpvLBuSsxlVOilUmjClPcwfUSitcVbqq1Tl6L6iFNTyRvTeipNUXAMLq57ItbU81ytVgz/PaP38W7X8/f9Tlp+AjM9ja/iOobf79BJm4jxTd4DmR+rZyL993KSZXkVYsETYsKffHG0GS9aAyzVSLCTM6RPgtT6bD2eVngUVRHUb0yUV283/i7J6Z/YqbXynJtaf1jKtIEZQgfU8SM08Zio5ZFTbOQvi9uqJVz2yL2wAxIJIpe2qZTvX2qUzm0cyo+pZOlRP0+Y2JVGh8p39wQsU7DUVpHab0iab1t96tLZjYhF3eR7Rv/tbxG+5RvkpH1xXVFTEAoQleaIDUebTwljN7V8spwCpXyHSeZnxg+iuooqtclqvMubwdIKKiEqQYzx9GM6NxF/W28C/aKT7KgFHUlOhUafeJOOWfr7R5P44uAEiwUSJp24fit6Qefm/n1V8m6kmytBJ93nj4vPzSVCFO4ouEMXXsNZ2h2symAtxUVZE+Ozdmdz+o1CEqVJRA6w8cGB3Vzu+gEEoax+zArIKpPoswlrtU2jJoYezgaubeBhG3B8xDihQTnYx19JHhl9UudUz2pKrmhwckMwzLFJRgmg/mbQUi1GyybKq0GjUMaF3chsZrhEiHJvS+HgwHth85y44hSZ0pBAWeUqfR7PzVFwu4cEsGkX4SJH5pF+GHM71iWAWbN2CueL+54vrirxRkSTf7ZeczJ/E1Hmngtc1uqhvY607Ycrg1odZ5LMvJFi50/4KsSnoc6Kw2lztG6GDZz0tKKbKNQM74R4bzaKaToolGw+EWz2IwNq9C+8QkfvBStS5aSXKP6RKu8ckHQJeWRlkzOSc2RQ/W2lY/lpmsdbKCmisdHPygV7U5bt0WRIw8x85El9GXYUvOaKFRVJgOPXi4ntN39bNVqxY8jwZkFmXTcDzaEfRL84QrshP8dF3t6p/NuxSPD7d3LkDwnxJ9gaKS+8T7mH9R6J8z8FyXl5LYW/Kl4yVLZia65z9SLNrgG9brAqzFfT1Ifp0WEwcbXozTFsh3V3gZpa7358+M9rUD/AT/nyfA= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all session(s) of a given agent. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-shields.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-shields.api.mdx new file mode 100644 index 0000000..767b5ae --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-shields.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-shields +title: "List all shields." +description: "List all shields." +sidebar_label: "List all shields." +hide_title: true +hide_table_of_contents: true +api: eJztWE1vGzcQ/SvEnlrAlhyjQRrf3DRtAySwYTmHwjGMWe5Iy5hLbvghWxD03zPD5VorWU4doBcD0kW7XA5n5s17/FoWDn1rjUdfnCyL46Mj/qvQS6faoKwpTopT8VH5MKkV6spf5O6j4qCQ1gQ0gS2gbbWSwBbjr57NloWXNTbAT2HRIg1ky68oAxm2zrboguqcVhCGvcA5WFAnFbDx/22tKgpBTRW6QV8fnDKzYpU6z6mLu6E8bXQSb1T1445PfO8atj8cFGhiU5xcFY2tUNO7TzjRw5yite6mKumZU/TIwXtp2fBmGo1M+B4UJRpZN+Bu6TlYq/PfzczZ2Hb5Nm0orhPgPrDz3keFU4h6q2lYu8saRZ+54NAPBOg7WHiRDThzcLAbaagqxQOBPt/A3Bo8m1LSDxYmak1DPbyXFD+CGTYRTCXVaNDygO1W6QcNOZLV9Wo7sV/O2i6yX8U7a6ZqFl1in0jZYEDnxdQ6EQiAPtXVUxlNQXs8ICl8i8phxeUc0GqTG5kJVI2gguYgJ7uhPxUephgW2f2gDDUEIcGIEkX0WIlgBUlF3oosqBGF+jOxJgENAtohVx6Rhvxtl747jnyL6IO4Ay+UmYNWlSD0GtAEYoPV/6d2HyDEIdsUDTtLMG+G9c/l5bnoehMwFSYRdhk+VuGm6XvnuPTcmfhOBbAuCB8bEtlC2GkiBaY+d7WStVBd0k6BCYk1VJzue/KTyBdA6ed67nqza20N5Sbq2IA5dAgVlBrFwGYjHPakSOJg5DOyHEjgVHy++JiT2WKWQ7LGOYrGOqQ0uZ6dUqC0MXQCaVES16WwUkbnaD7Crbh+ho65wn21HsAbMDTB9AR4/YIkpjTvpRhOzz+MxIWNs1ovqDxa2zsvLv56J978fvRmxKDhPTRtx4yeX0T1tb8/oBIXHcXX8TyT+Vk5x293K0dqRaoQNZl7fqCpm4zNoh+YRyUizNQc6b+xkfowtqrBvaj2onphojp+u/Z3SUz/xEzPyvJDaf1rI2mCEMJ7iVgxbCw2CCi0alQYiXPaIVBAgdgDMyCRaProRllvr59aqTy6ORWf4GQpIcHDxIoG7wnvQG+YYdhLay+tFySt18P16gMzmyIXk47ta/9ZXqePKd+DUY3EWSQmIDRpVSqRFh5j02a26uX1cHB4dNj7wcB7Ue1F9bJEdbTL2zMklFTCVIOZ52zyaYqj34yWj1p0otb5kOf5WmR9AiVTsqDn2hIoxSzdALQQanoZz1+NsxFfCySd+3Smjo7Dq0NoT8ZjWl8Pa+spxEOtoYFDSk3ejqRt6FTMdhSwCotk+Gcv6qtr/sZlTgTqj4VsLyZsLya52JDvINh5l9H81Y5LBNaG32AJDHk7HDmdbvOB2X8xYutHQsS0AUZTtZZk70lnaXOgHCncdYWviOqC5wQ3BYm+GxQcfjFcPOsSqx8PXvJGXhPPHcz6qD7xvYxPBGkJR5IgY2KoQk+UbyP35XrG29k5izPgfRi3mnYx7CcVcJkLfdUh6tcM4nJy+3JZ0gbos9OrFTfTHsotOsakWajkol3xzUdNUwZfb1GJb3HB1JUSW+YSHVdiujjZnotXQ979/f6SCP0dVpv0tw== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all shields. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-tools-in-the-runtime.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-tools-in-the-runtime.api.mdx new file mode 100644 index 0000000..33a4332 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-tools-in-the-runtime.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-tools-in-the-runtime +title: "List all tools in the runtime." +description: "List all tools in the runtime." +sidebar_label: "List all tools in the runtime." +hide_title: true +hide_table_of_contents: true +api: eJztWFFv2zYQ/iuEnlogsdNgRde8ZU23BWgWw3ExDG4Q0NLZYkuRGkk5MQz/992Rkk1bTmwDw4YAzkssibz7ePd9R/LmiQFbamXBJhfz5PzsjP5lYFMjSie0Si6SS/ZFWDfQWl7B2Pbr8Z3kJEm1cqAcTeFlKUXKaUr3u6V588SmORScfrlZCWhJj75D6nBiaXQJxongNeMuHsWN4TMcJBwUdvdsxQuIRllnhJrgqPVF/IGjmB4zlwNzuJRksTFil4E3t/4Xl2/Z71XB1akBnvGRBBaNIw+POXdLNyzTiBJ9ldwgAgfG/pcrXXpdg/AQDOwyNMCvLUPsDXQmnRMW5pwwgQyYgHl7cDz3DOLKcRNJA39XwkAWuRhhpIEr72PMK4mMdKaCTZd/5oAmDdoVNjKMD41NNtYm5E2oqQ50Diurzc4TreB2nFwMl85VJYlNLTDRK1UVI5+DzaBEbwIXohc1CRb3LzDxKuBiUy4rCODXYibGTGnHkEZTkWHIFrgWnmUizO9F7BpzaSEO7jDQrUWbdTTRDMTphJMEnWpFb0m9Tfy9VUphLJQH47FzH/pOi0rxiqkUEUFWegrp9EnjaQql8yzBb3yjrixVtT0C/0dqX1rp5RIlaxbD+EhXLqphB6dzI0nIn7bsfdlaZaayqAuhmKmUE1hafNF/cnZLnprkuHUT9jCgPm0R0G27Twt188HD4+hYTZBP8hlAiH1BmH7att8NMLyEB3DuI7dUCrgUGUOGFlwiUQvE+a9tftZxV8XFv66o7YI5GPRYGI2LzIDCX4doV639bAyVBhp8glGxuTaO2aoouJk1FR78mMdcpDlVRFq0EVy5oExVf/d+fN4xyHJfz2E0uZZa4dpY/mLxX8IhT0LhmlW6xypj7bCv/S/1YlIEP4JAY6cxszgbpsAKbQCXSfn0mYvEZUtIxVikTKdpZQyg+w1ch/C5znCTrWXwIor7MD0TvOaAxsZGFx7DZe+6w/q6muRyhumRUj9a1v/1E/vw89kHL0t44kUZmNHwC6m+8vcLz1g/UHyFZ0/m18o5/7hdOakUqAqW43RLP1B6OFnNGsNkFYkwEVPA/4XGquJjK/xmcxTVUVSvSVTnH9c2VHZDTK+VZWNp/aUr1ARGCJ5SgIzCRmLjDnCTKoTrsB4eLRCQQ/bwCe5gTOJH0+xU75/bqSyYKSYfw0lSAjrEYm4qBU8Yb4dPUIfhKK2jtF6RtN7H+9U1MZtOwneB7Sv/tbwu25RvgpF12G2FTABe+F1pBLjx4K0IA4bfanlFF7yN5scLho+iOorqdYnqbJu3PSTkVUJU4xNLq6HrWD/cB5NWf8LfArmU/tblj3t+qwvDO8l6L2zYtLUSGv0wMboqHwRJS5At3EjNrH1FRXvXV3E7j/mJlFp/5QueMauduEPR5KWlyubKTlf0Gk2Rlg+gslKjEneDufnUY81oAlHZVTdmBW8/MM+UiMqIPVp26I9IHz4zj4d+ICSfA7C6Miijg6hLniOaov2W3yA1A2MIOkF/8FSX7LpEUqW9932ZXKPZZAJ+jdzl+NCdvutSnE5rlnQpi6c+izgoHHECWSpDXM2dKy+6XTxsnebaIl9PpeQFP0Wepz86qS5828wCsle4mZ941VT44T19I837eDZNBprP7mg+u6uVz+sWFzkP65y+2xJyKpR2rWTwuIjFlh21NetunP2m2MYfVmXwt6GGSxaLrj8pCoNxNKEKZJRR2iDMmKdgg1Fu4Bs1PqQ2vsS1jY/oViex6Bk+aVDd4BYiQyenxDhiPV71mXfreKPfu9wLd8+sWUx9pG4p8bBLCHxq5zUlhiHWMSnwMaIF5pBSTyPn8xGenL8auVjQ6yBTynkmLG0tK7E9i/iwirIV/g+YbalivjOL73377zA8+xeVF/Bs1LEVnHt6wM19RPQfUicyx50YjxwUuDD30rdTo1mtI86aon/7PECN/wP5rjYW +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all tools in the runtime. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-vector-databases.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-vector-databases.api.mdx new file mode 100644 index 0000000..819f9a2 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-vector-databases.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-vector-databases +title: "List all vector databases." +description: "List all vector databases." +sidebar_label: "List all vector databases." +hide_title: true +hide_table_of_contents: true +api: eJztWN9v2zYQ/lcIvXQDEjsNVnTNW7p0W4F2DRK3wJAGxok6W2woUiUpJ4bh/313pGTLsZOmwF4C2C/WDx7v7rvvO4pcZA59bY1Hn50ssuOjI/4r0Eun6qCsyU6yU/FB+fAFZbDu7K2/aA0G2UEmrQloAttAXWslgW2G3zwbLjIvS6yAr8K8RprK5t9oGjKsna3RBZXcFhD6o8A5mNMgFbDyP7ZWBYWgJgpdb6wPTplptoyDZzTEjSlT2ziJY1U8PvCB9+nB/RcHGZqmyk6ussoWqOnelwp1QRezCNm4yOmaU/TIwXtp2XA8aYyMCB9kORpZVuBu6DpYq9u/8dTZpk75VnXIriPgnuDenBon0Ojtp/0ajihmYSeiw+BAgL6FuRcvVjYvxMQ6kW4Fh5sDs4ISxyrHouCYU4o7MNj09g9U0VsoUayMRTQWwYrGY9/XFA26SJxNZ4Wq0HiVqNQ6VMS3KRX6vsezbuy22+QlJrLKdWwowh1FpjFAVjwp6PMeyyagPR6QWL43ymHB5e7RbpM7LVO2cdudHFU1qKA5jE5jW+l92azKqowRRh8ioQSYQnxv0M3XWa9h8AMGYHNWljXjtV3zn8EhireXxM5msZVR90JMiNtCk00v6lUkFPSSg/ltV18alQwEZUxZ3IIXysxAq0LQBBVoQqbC4v/rUT5AaPwTqPj3aHQu0mghqe6xdSRsfqSbd85R8HEwCVT40rogfFNRa5iviB3H3JZKlkKlpJ0CEyIVwLTvo59Y8ADqCYpNntNodq2todxE2VRgDh1CAblG0bPZCIc9KWpMYOQTsvzlU52I9as4FZ8vPrTJSAo+R24OBTcJh2SNM6S24ZDS5HrGygnIbROic1+jJAVKYaVsnKMuivfi+hkitxXuqrUCr8ftCNMD4LkNRnMMp+fvB+LCNtNSz6k8WttbLy7+/EO8/v3oddQj3kFVJ2Z0/CKqr/29hUJcJIqv43ki81vlHL/ZrRypFalClGTu+YIWHDI2825inpWIMFUzpP/KNib2ikB9ay+qvaiemaiO36z9jYjpH5nprbJ8X1r/2oY0QQjhnUQsGDYWGwSkRapSYSDONfISHIg9MAUSiaaXrlupXj20Unl0Myo+wclSQoKHidUYvCO8A91hC8NeWntpPSNpveqvV++Z2RS5uExsX/tv5XW6TfkOjGIgPjXEBIQqrko50sJjbPzELjp5rfY6W5vURybei2ovquclqqNd3p4goagSphpMPWez2odx/Ds2f6D1jj0XsR4cbZBJyzwLmdJ1aQmfbBqPMGoIJd0MZy+HyfqwyBmztMaxzSJrHAdbhlCfDIe02h6W1lPAh1pDBYeUqLwZSFtly2u2o/BVmEfDs07iV9f8jose6dRtL9leXLK9uGxLD+05CjtP2c1ebh+BlKwUv8EZ6LO4P3MoIYh2U++/GnHvR7LE+DmMpqgtNQEfd980i3Kkd5doEI8fuEO4CUj0aVJw+NVwKa2LHN+ePOfPek2sdzDtovrIBwhpD18TjiRIxiSdYjxeyg0QFutG+LhVK96Ad2FYa/rKYc+xpIu2+lcJ4179qVhcY361WPBMn51eLvlxPJVITIqNKudKUnnJgLoKn9tR3W9wzuyWEmvmGO1omngYeL9dL/t8/OvdiDj/HxqUR5U= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all vector databases. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-all-versions-of-a-specific-prompt.api.mdx b/versioned_docs/version-v0.2.23/api/list-all-versions-of-a-specific-prompt.api.mdx new file mode 100644 index 0000000..1465377 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-all-versions-of-a-specific-prompt.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-all-versions-of-a-specific-prompt +title: "List all versions of a specific prompt." +description: "List all versions of a specific prompt." +sidebar_label: "List all versions of a specific prompt." +hide_title: true +hide_table_of_contents: true +api: eJztWF1vEzkU/SvWvFCktmkrEFCtVgosu1sJRFVapFWpkDNzkzF47MH2pI2i/HfO9XiSSROWsNoXpPLCNHPnfhyfc23feebI19Z48tnpPDs5OuL/CvK5U3VQ1mSn2VC8UT6cO1vVwV8kc5FbE6QyykyE1FpMyXmYe2HHIpQk6mh+mO1nbEgmsF9Z11rlkv0OPnt2Ps98XlIl+SnMakI4O/pMecCHcFGTC6pNrZChbyWdkzMYqUCV//HXbTo9Ox8cUofdeq2XSN3PPJymCkSguyBuVSjFVDolRxq1aZlTaXWBmg/Fh/SzF9KRsEbPhG/q2rpAhbgtyYjGM0qMSoeeF8Pzs8NssZ8l3HqZKcA1IbeR2ofWUuwlA+GDRH2MfxDH+0KZ3FEFpBEWZl5O6TFHaOv4pIofV39l1NeGhCrgRY0VYoytq2Rgl9KLRzU7+u3J84NCTVQ4KKUvf38Uq+gw2GGFUuzF/eBMMmZPwn2JtpEV8AolqsylESMCnshHmR7PsEoVViUQu1X+U0Fj2ej+eo+s1SRNLDm9G0vt6X4WL1s7uC8iUwEv1hCRHMIp3/FcKB/DJ2fLn4FXa5cIt0A+sigUe5f6vMfJFN3R10Y5wuJcL7nQX7M+tmul3exnQQXNtbXS3FjNYQcOJG4blxMeajzz4jJrQCCLyOJdTWZ4Jl7BFgVHgq+Dyli/0bKS4n2Q+Rfw9qfKisrtpbulm2zkvmwzlS1Ii2CFZnq0ifmYAVJ4sq1dsYQ5PMH+FqxVZiq1giacqKRmQlPx/7UliDA0fgf5/n15eS5aa/TOIjI1IfIjWb52jmnFxvu8bCV6C3pMVUk36/otRZvbUuUlc5OLBm1MiIwEn9v3MU5UHnq33jVya82htTXcecqmkubAkSyiQnvfrKUTxWhQs8l3qHLvXd3S6bEYiquLN6mYvuZBA0f4mqZMDMfEbBsUR5Yj24QY3NeUo33lwuZ54xwh/L28foa+aYW71VqC12N0hOk74HX7qxiDuzEHbv3iwjaTElvF2Gptb724+POVePb86FncFOhOQnnU5xeovor3UhbYSiLFV/nsyPyknJMX25WTawVVCLR2wY0CkFt8bGadY/YKIkzUFBubrGxjYtMOqqIHUT2I6hcT1cmLVbxLMP0tMz0py/el9Y9toAkgRHc5UcGwsdh4c9SqUuFQnOPYgIQC2CMnOBgL3jldt1M9/d5O5clh2xeAk6VEvB9jbRpDd8Cbz12UYHiQ1oO0fiFpPe3vV2fMbGQu3rdsX8VP8hpuUr4DozgU7xowgWQVd6URYeMxNt5Pik5evfP+vTPwvzh+ENWDqH4tUR1ti7aDhKJKmGpy4rmadPvi7Lfcwe+Pc+QKy9VUp5YO93LImh3OM76k4/v+vVWxQ9woy61Dlt6MYW1itLzqLTPA4nLA1WoE12B1NrTZzRUWqAqZlRam2YSiWjmL02wwPR6kG+Rgvkx1MehCwbLdjtuaGse4liHUp4MBDgYHpfXA9kDzVfjAx6twbquMA3oC0irM4od/dN3o+obfMT9jnt39d3WVFu8TspG9WW8clE2Pt+DGovZr9JZ9wfU9x4EJipwCaf/RiHv/sKgUT+5kitqiX3k0iHiqUQ6tybWMLXhQwM3MjWXeTWGko488+9NxehDspvMR30A0BOrkpMvqLV/kfWR2DRzROxiTxJvdebeGyHzVwH/CReILD/YGtcZRjXOKiz1PTLlu0U9cwdPp2kCmowvWlinB9vP5CKe/K6cXC/4ZB0jHXLhZjW8iMwrl+bnYPnzql7N3kfj+WPwnuWwtspvMmVmcK+mG/8LjF5qtiXdxA+MSnR1bGKfdvh/mOcUxU/flxpa5Jr2/Xl9CjN8AanfbTA== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List all versions of a specific prompt. + + + + + + + + + + + + +'"},"variables":{"type":"array","items":{"type":"string"},"description":"List of prompt variable names that can be used in the prompt template"},"is_default":{"type":"boolean","default":false,"description":"Boolean indicating whether this version is the default version for this prompt"}},"additionalProperties":false,"required":["version","prompt_id","variables","is_default"],"title":"Prompt","description":"A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."}}},"additionalProperties":false,"required":["data"],"title":"ListPromptsResponse","description":"Response model to list prompts."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/list-files-in-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/list-files-in-a-vector-store.api.mdx new file mode 100644 index 0000000..47d10e7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-files-in-a-vector-store.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-files-in-a-vector-store +title: "List files in a vector store." +description: "List files in a vector store." +sidebar_label: "List files in a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztWltv2zgW/iuEn3aAXJyiszPNW7bT7hSbTook3cWiKRxaoi22EqkhKSeeIP99zzmkJMqSHbvtzqCACwS1ZfJcv3O1H0ZG2FIrK+zo9GH0bDzG/1JhEyNLJ7UanY7O2L9F4rS5gj9xLq17LXNhL8M9lmjluFRSzZnLBMvhANMzNsNDR6ODEX4ulEO6vCxzmXCke/zJIvGHkU0yUXB85ZalAHZ6+gnYwcXS6FIYJ71o4XF7zjoDPOFcKma8yuGjEfKmB7H4F3SR4S0mU5BEzqQwB4znd3xp2Q3duhmNHuEid7Ek3Bi+BHrSicI+LaFMh6WLhXmv5O9VLAebaUN2Q3uhDNvouSB/TCw65Igu7qx0j4S3AHfAbVo5MagvT1OJDHj+rusbJS5mo9MPzQ1V5TmQa95Ptc4FV/EjVRVTYeInQdHoiXdA9CBI8vjxcVXhf4nl4YLnYNtWBcat1YnkTqTsTrqsY+ckq9Rn4AcmMHBivuwD/yp8wioLFNBRFvDrXA11JMWkcpoRMQsW6lliDVb8p30fQ7BY9DCvnO64vHkwKCCSiZyLh29GJDG+LCDiElYrPHp8XOfJGc+tOICU8HsljQA4f/ACfjwYOelylDRKBS8DwVqKsyERz3r8WW1wks/jkBEO65zRd/iu5rMOWXYMGD3axoT+eDCifxNZ8KAm18PMS61mcl4ZSnKs5IYXwgljmzhfodVYAyTbrDPdmOiFMDkvJ05/FioOUsChmENARTo/H49X1f2Nog7zsycA/7FAkk2FuxNCMZ5+4gmkjBrVoG3B7yeev5V/iK2Y/9xn/pbfy6IqmFoVArT0zA5YUUH5mIpGmJPxmHGVsufjF3/fDbmD9lqnynYIvyLfeRd/QRg1qNmF2wBi14QTpbikxt8UclMLPwgqYJpKJFRIxYEfui8gbPkbHAPSQcoC6jTCHCs2xvSTlW+fzTZmsyhZ7NPaPq3t09q3TGso7xZcqeUzAnvBCXdrfBzLcy0LAXoVJbvLwGFNv3fHoa9MU2gIAWP4NI555JJz6ybCGJ9hu0T/dlF6u/7AXuEJ6B4hgAofVdIPTQziIxHWohFmHB6kT8eQTsVgH97PIMIAcoJ4/d67PYhWm+RQK0CX+0QIUNiXkI4+XglkD5qkNNqF3pgmDxwDQYPKGwbsafl8MMF1qf5aFVwdgrNScjcJy8Jt5o9Omx68pr9TDKG9WoGGcYtT7jn48lWwFaG8sgO5sTIG80rkN3+UUkE9cgzMBj2jJ7oocwB2utE1Uk2A1RwGd7vxXMJVIvL8CWoBYo/rjXDl9QYLVGiuyXTZHQ8HsmQvRyIx9J4tIQf7eWoKxSmTtp6imCcLXOLRdLLNUP3ml9rSnfrbWUwERrvhRGLotdNvOxsPjY+d/NKgpWu0vnLrzX5R812Z7Euhzt6ElQyj0wyPMy8n9RvdG+dhI9NrTsIVMvpMGkhb21g7ymGt4el648pmD4TdQcnn2HDi5To7fjEbvL0Nl4zbSYH5uGVTryAilAaPd/n+JxNAmJoaMBOHv6Jp5hhfQLRQUpqKpYb6TcAqMYnsBKwGU7RxiuQdhkNv6dazVrONmxldkFWofJDQYCrecT5VTRD3+dC27xqsiqJC8aNaJ9WC5zJlmIJ5jtWKCtI32uq1OfWJcvzr9fW7Oq1S+m7L/lNI8kWKDkMnzGymjWO2KgpuljWwfI25y2SSMemVNpIrjy2uwufEh+ILEku+LWd/GlnnWoFuLOvWt+hORxzkJCFJYxrfKV7O2PvL86AMFAHsOynjQsNiBNwWiwDquPvgU10538qXIpEz6MV0klBpS8SKXLtAvcmD3luN8SKk+wo7bDzTATbKcPbuzRG71NU8y3HoynN9Z9nl65fsp5/HP1H6E/cc62iML5oZan7/4Cm79BBv5dkS+SFynr0Yjpwkl9gKQDwziy+c1nBZLWvCIRrncoETSaErRanZQbO5D6p9UH1nQfXsRcvvGpD+FpEeIsvGofVfXUFMgIXqWYKYYuPEaMY4Yu+gOINADtDD59C1QamHAa+uVD+uq1R+nmFgTgwlqNgpAqtS4h7sjXt/EcywD619aH1HofVjXK/eILJBcnbl0d7yD+F11od8bYz0iF1UgATBC6pKU9w+KU3fB6Z1eDUdce+r3w2E90G1D6rvK6jGQ9y2CCGKEoQan1vUxk9mby5GvX0Yjdprpy7AfbO5pD2Q8l+7rO4E0OPwuOQu668lQfk1+w7wbN7yR1Mhy9YNzlTgll5Q1l+745IoCEQ1uRYDqrlZbkYanWfaj+TtGjysF1CwKbbWrjIKE9I5HUdQGg6x0C7FaSV+AqmPXiCxkJkwHJ+Nu+oEmPX0qZNHrJA2KaWT7RS6wtxAV/ySSrDbdrVzS8263wwHLwQ9j9gtt8mtzxY2ESrFGdzTQX1ukaH/GF/Fn2+n2YCn+MztoNgZA0Bb7X90AtkAQdouTlB+JHeL1gbneLUQbS7jDl0hFQBrqaGelDnu8aIlzBdrMBUzXHx8MxU8vT9XB4i4Xdzwmo4jtmiVFQoafmWl8mWIkxDFzc9WQj6G3OSPPyHr975rxiOQKDMNyo3mgnoGTIeno+PFyTHUGsUlvopTpz1+WMmkj8dkRdzGUufks25lMOFnzpWnx8cwsRxm2oIdDvOcF/wQzJt8PgLrkJhWANykW9LFX+o26cNH/AwLJ+EiqHKO99kV3mdXoXz6jSSufY31OFicDOR07DZsp+7yuBOIKROOwScLCURuFFv5B62NoJUCpJdSQy60dS6V+EWN8aWUMg8mSjODGLCeKDfiBjfmuTbUJ/SJT3E1kkPnQLt8L9Vb6MNy/zV3CXaEpgZtEgLjqXrYscND208+eTEgzIl7dwxxLGnrS459CED54C3toYIeOFnZviMsTvul1yMG3IuoQDIPD1OYTN+b/PERH/voRjik0mLrlg4vkmNt/qpiOWilz2IZlXj6zRy8p7j9Io3+mmq5QbW62H+9av/nerlBh7qu/wk6fGXB3KBEU9q/XotvVTI3SNsU8Vbaj/gGRsAp5vfdIv4y1OYf2M4d+6CM9S9i1TKWsJZ9NY3RD2QzmC1BIRTcnzpLElHGQd8b2jtl95+vrqEQ/w+Zxzp5 +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List files in a vector store. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-input-items-for-a-given-open-ai-response.api.mdx b/versioned_docs/version-v0.2.23/api/list-input-items-for-a-given-open-ai-response.api.mdx new file mode 100644 index 0000000..1ff7513 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-input-items-for-a-given-open-ai-response.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-input-items-for-a-given-open-ai-response +title: "List input items for a given OpenAI response." +description: "List input items for a given OpenAI response." +sidebar_label: "List input items for a given OpenAI response." +hide_title: true +hide_table_of_contents: true +api: eJztXFlv3DgS/iuEnhKgYzvBDmYnb94cu95NYsN2MFjEgcGW2N2c6BpSst0w/N+nqkhKVLfUOuzJThadF9sSWazjq4NFKveBEjrPUi108Po+eHV0hD8ioUMl80JmafA6OE7ZB6mL01ykxyfndvhJmpfFSSGSg2AWhFlaiLTAqTzPYxlynHr4m8b594EOVyLh+FuxzgVQzOa/ibCAibnKcqEKaVaPeOGP4krxNQySsAq9z1Jxughef+mjIyOPii6UTJcwpinU51T+XgomI+BbLqRQbJEpVqykZkWWxSzkcRw8zGA6L0rdT+9NqRSQYmY8yxZAS7BbMWdacBWuGDJIakGqhtY2TVCkBjUGMO/azLsmRnC1BS/jrnc+J5eOfYbUPQlnjMe3fK3Z1SaNqyB4AK54FEkkwuMzT58LHmsxA5z8XkolQLVfUMGVZqwwX+GnLGKUpgmU07IApHwUWvOl+FXML2hZZPJNG/e/1hqr7MAyosESQ4QsZRZhFXoPQK3fExcwSFmyXXBtLvSwuRL6FOLESmsJMnEnwrIAPT8afAsZi0now4md8Gt9ORZ/m0QQgAgxDWsM0WmHgXkBosxBe61j2+HdElbSkmxc/T0HiQRP/UdpmcyF8p9UZt5gfhuWD1+3sPDsNDecPWf/EesXNzwGDNbSMK51FkoOsGC3slhV5kWtkS6nQdsHCsZwLlOYSM+MLRz9lCetiGku8AlGDaOpw0z5BK06Nwmei1jc8DQEGOOE2hUtqg1B9mwuilshUnbEeBqxl88J5OKu6Gf5EkYxm74c6w3i4wKjB8DaMJ4KneSWvWFB8z3MbkbNc+som9Jc+Jxr+FmUKgXMzNfdAeFgOzJ5aJxCcUIucdF0WlbZVtCWZt57jD55XkFaEz3QIBqWLtMQB1X5ZYLHORJzgQ6HhEwW4WpZJrDigETy74vTT8y823TeinxNbkgqsbNa8sjWm9FJxKdgMsgQI3joPq5A2maTSXWgR70jK/uWttl+tNM4xFmc+DYe5zmWmW6/cdx+P6cBPZglWl+NAty1JdSJOzegAzo77D7RXBVDu41Eu6tN45zWwjRcBROiEjloHwFAILM2AsjxDcAVK16wpYBxORQVGMl5+A2sS9OSLBLxIOPVG76t6qmrEJqwkRuYw9FXW/K4TD2cDsEOTbimRX3EbDxuFN52xe4gVU8eu8myCw4AinVmy8xlG6ONOgddtqEcDX9uOjGOSngxBA2RgEQR7wJDreM4u22pmuv3K7lc7RzAyyKDEto3ED3a2l6JGxEjFAx3RugEIxYwH4LcQHYGPoH5EuwEfF0FM/gFGcDfYPgVUTaZZRh2aIEW8NTPp6CHZtsEh79elyoelYk+n3+onIJU4Px3FCCtmSdA8qRV/BOflbGgRABIpJbIlBeZQn1YVK4/Ue1k2JwFCc9z1A82AGpP7t1L7mPPXyL2NByov4O4j0T7SPSdIxFyPnh9lHNqd6s9Ig1AhKkGt4OL/xzVn6ZZQTv4nS24wcXbkC5jKM2K2y3GxpvmcYRjs6/B6EgYwD6+T6bEQsCOLoSq2W++jd+qtxCSaSTuPCoS4LJsaYmdZZp8hNEER9BJSt1BACs9m+JZNme2tq4Mi53eVtsFmy1vusxHnZiK3xpz5HZOMdh00LkIQf0hqYhcsPI9K9qQ2rQPhRDA2kG49WI0Bn0KBoIija6HGvpdGrHcGXvTzDrniICmnWmzqorBK1zg6NFrWOP3FkY4rAXveCIGdsxKFRLuByUQL2t0kpqA8tocTdU5IQ13AwDfCXZkfAjWIQILhW0oX6ZBbZQ+fNsunlDX3fG2a8zDzHvV0R/ZBeldQXdn7OyF8hRrN0RpGr+9U+9xMAQDjnwj+j3egMRbzovVdoqkpz1KfkoN1moalQrOiM1JG7YmHvsrpH3JsS85/vSSo1kaPBqU+wpkX4H8lSuQzvLg0cjf1yb/49rEKyOeJrf+H9Qp3e2cY68/8jC5ndrssww7prXNI/OMeqwPdIlIZSYQ9jc79VoXItnZzoywTYli7BxV6p4BXGsJ8EztPac+2LgWuQ8ar23+p57K2pRi9djXTfzocdrsnSpTJUTanaPecCWzUjM7hTK0dpnM0dPs+OzkgF2uxJpxJRiezpaAV6j9UjOFzUXIQeHb09hS3gg66k2YOdUVTGO1d0VCXAWMrpDN6D4UvATyUrM0Y1AYEJPCTILZOaZgYC0ETUGMYDqEIAPsa3PIsqOz2X2l0XRPTccQj/NNRNmBghgmNiBQPfDJnxKdHfUOzhp71EGXjisePYl33HnebvGj4FUQb7sT4euEmsbA49/abltfUokBuwGgeMsRNWBKGeEpQMJj7DqLKHiyK9dbjtRVqv3r8vLM3WQJs0iMKMTeKYU3anAw2IrpVQaI02WScLV2RZWgMbcrGa4QqSi0ktw246EKNO9pHUKdO+YZsrIZjUvHWQqysVWZ8PSFEjzicygRvTkNdkxawlgWDpDSv1ZExxxGGHusU+JVC3A8JWA2BFqW4H1GmZpTBKrN5hnggxzZ7UKyMKQrRKHY4GsMvusLdbaitMrzkE5q6lBefd6hsoR4oKh1npXLVbwG88RxdqvZ+fs37Oe/H/1MBaO440lukOHwBVCv1/sHjyCcEcRrfgYi33rOq1/aPSeMJR7hrDjeEcUDrSyDyenaEaYozCl8ws8kK+2RrUzE3qn2TvWDOdWrX+r1LgHpHxHp1rO071r/zUrwCdCQuAuFiERkdrK8ECyWiSwO2FksOFYagB6+xDoghpfKZaqfujIVVIM3YHxQJ7oS1BkRAqtMxR3oG6+rC6uGvWvtXesHcq2f/Hx1ktreyIVBe72+da/jbcg7ZUQH7LQEJAieUFaa43cKsP/D0jVy7lXVvS2f3nUS3jvV3ql+LKc6alttgAuRlyDU+FKjNMdLuun+tXUX5u2zDG5ssbexH8PPVmHrC3tm8G1NvRPT76s+iHUdJHhUtbU209/J27orbFXk23eDF1yztkqhYIO+7aN1J6PiiC8KckbiBT+VWbccB9AyyBAwEJMiaFmaOzPQQ3XkfInnbxI//PG5saAZws5cLMw3TFP4MZOfliGZhnEZ9XFUf+sBoTc2zRo707VmfHD0c9L3wafPIlVZPQyaUoxlhhfzSRzCy0Rt4ncuqk+wDtgHGo6xRXEIacx9BffSfAUHGcy1f5hNMBhVXx0Nk85rwFZSZCrqAyJ6BQ2zjgC8ehdiDQZkesDe1hwhgVG2xwZ5mdD3djq0HDT6dpbNjcMmzDiGNQ93gEL/sxVseUFIWGXARbAUlCupPR4c3rw8hBibcom/VXMO77148XBor0kSIkAAqhdMdKFjpmBVFPnrw0Oo01+sMg2rv4hjnvAXECLDbwdhllDTTQsIfLJY08S3rjj48hXfYboglbg2Fc5nFzifXdik4Y5RcHEj/M3LFkthjtWNbMP9/OdTps9GoGq4kUDkKmUb/yChC9pIizTKM4mfoVjoSQWVgmuQ4kET4koteEjtSyDKlbjCvlmcKcqO28Tn2BCIIZ4q7KUarj7ilyrmeCoHPUIqR51YmI7NAw29eB+3jCZkYYrt/sM8hl2Ud75oYPTFWMIACS30koDv/guGWfB6M/3UeALjI2aQyP39HHZrn1X88ICPjSsiWCKpsZypfahTtokZo1XEb2LtpSlqPcPfdEzwBAz1pIwdHFWZaipLI3PGDlbqHDWVl++ZHnYI4jLZRDEekR92MOUSU83UV/wDdg9zDJKjHOPZuc1Dz9m0Eq+VT1czpGufTce/7/R0bLaCfQlIhJybEcdhKHJf71sbvkbq+ue7S0hmfwADzZhe +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List input items for a given OpenAI response. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-models-using-the-open-ai-api.api.mdx b/versioned_docs/version-v0.2.23/api/list-models-using-the-open-ai-api.api.mdx new file mode 100644 index 0000000..98f75d9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-models-using-the-open-ai-api.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-models-using-the-open-ai-api +title: "List models using the OpenAI API." +description: "List models using the OpenAI API." +sidebar_label: "List models using the OpenAI API." +hide_title: true +hide_table_of_contents: true +api: eJztWF1v2zYU/SuEnjbAsdNgRde8ZVu3BUiQIHEfhjQYaOraYkuRGkk5MQz/955LyrEcp5sL7CWA/WLJ5P085/DDy8JTaJwNFIrTZXFyfMxfJQXldRO1s8VpcSauGrJn5xc6xEtXkgk3nc2wGBTK2Ug2splsGqOVZLPR58C2yyKoimrJT3HRELy5yWdSEYaNdw35qHPkUsb+LOm9XGCSjlSH/7bWZW9OiF7bWbEarGfvDKWsAwaKmuvBe0lT2ZrNLzBWnmSkvmONSmfkk+cHS+Xfk8ULYTEqy1JzF6S57qU5lSbQAA3/p9WeHd9x3oNNTeuIPff3gyLqaNh/BuHyKeFtiFLaYupd3aE1RCbfk0oCYCfcLubsFn5/eokp44oE+6QQxYMMQtu5NLoUzotamqnzdaruf6JMiDK24QV8nnfnz/H4WuTZQqEYBrArc5cZ26YfvEfyafJASBEq56MIbV1LvxBuKiIqpjTnodKqEjoX7bW0UaBgIW03nuKs2H+U2uwbOc/m0MZZ1Caqtpb2CEQp5cSQ6NlspcORNCgurdqjyh+umsyRH8WZ+Hhz0RWjkPyERBuoFNEBWVjTnMA1TyiT8UzICTlxbUzBQ0NKT7USTqnWe0L4Z3l9Dyc7hNdoPTWvR9PUpm80b720ZVlwDmfX50Nx49pZZRaAxxj3EMTN77+Kdz8fvxty0+hR1k1mxppfoPom3i+yFDeZ4pt89mR+p5yT9y8rRxkNVYgK5oEfonMwtou1Y/YKIsz0nPBduxZzuLe6poOoDqJ6ZaI6eb+JNwbTL5npnbJCX1p/uRaaQIfoURGV3DYWG3ZKYXSt41BcG5JIKII9ciYhEoNBP+z09vZbO1UgPwf4aCdLidAeJlZr6RH9xj7cNfcgrYO0XpW03vb3q3NmNjIXt5ntm/idvM52Kb9uRjkUVy2YQLJOu9KEsPFYF9EwjHXyejo571wb/sXxQVQHUb0uUR2/FG0PCSWVMNXkLHA1+TLFyW8nyxetfIkL6DXwSanmS1jKmJkvvawJemZP8IDnyqFHxYySMGSs8DKavxmhoVZqfsouMZp3O7ZcFq3ntKsYm9PRCPvuUeUCUj8yRtbyCCWrL0Pl6mJ1z3YoRMdFMvxtLfa7ex5j+BOxus5csL24ZXtx25EgkQPxOXgudf5mB6pxxZoJW+yRfT73PcdKRoFVYK7h5JMVzz4QKKWDMdmycVgOAvSXDg3aQ/k+E6LkFvNa4adSUchOpadPlkF1PrF91/mED/gG/Pdyts4qQ5qI06CPkCb3xAKqPXHd6sVyszLuZdyJOtJjHDUGpx/OIwG87BhxlzueOcFI8Fv9xEPGnictlxOcoj56s1rxzziI+UXmWVrKJowwYIcB1h0ssMyHL7Rg/itFDTMQd542/XvzfEFf9dn6x4cxVPEVs+ptEQ== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List models using the OpenAI API. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-tool-groups-with-optional-provider.api.mdx b/versioned_docs/version-v0.2.23/api/list-tool-groups-with-optional-provider.api.mdx new file mode 100644 index 0000000..21a53af --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-tool-groups-with-optional-provider.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-tool-groups-with-optional-provider +title: "List tool groups with optional provider." +description: "List tool groups with optional provider." +sidebar_label: "List tool groups with optional provider." +hide_title: true +hide_table_of_contents: true +api: eJztWNtu2zgQ/RVCL90FHDsNtug2b9ndbrdAiwSJ+7BIA4OWxhYbSlRJyolh+N/3DCnZ8iU3YF8COC+RJQ5n5sw5w8siseQqUzpyyekiOTk+5n8ZudSqyitTJqfJmfiinB8aoz9ZU1fusrHoJ70kNaWn0rORrCqtUslGgx+OLReJS3MqJD/5eUWYy4x/UOphWFlTkfUq+s2k746S1so5BilPhXvaWmUIQU0U2c5Y560qp8kyDJ5hiB0hVVPblEYqe3zgA9/ji+0PvYTKukhOr5PCZKTx2+WKdIaHGaI1dpSN8cwpOuLgXWrYcDSpyzRA3EvGVKZ5Ie0tnj2Abv6Npgx4zLeofHITAHeAe/NzRhNZ6z2vu2UcImphJqJFoSekvpNzJ96sjd5wmkVajajMKqNiYR8Hv7ZqHyZbvnMS3y6/iPhZhKn5wRvh8akNKVnCvcwyxWZSX3QcTaR21ANbf9bKUsZws2cg4pXX7Brz7/g9C14tTcgCYWJ/dO/JYnLRULfPKUs73cuz/bEsElPS+QQxrCzKWmtMtPo9BqQky+4rkGQMhnberJi1RfzOiyaS5c1yO7VfzqsY2a/ibBWlQCJ1gaycmBgbsOXiisiIF6HbEdWmMhoddJBftYY9+AfPkXZaespCPE4UspTT8GtKiNKGKmzactNhw3UC7mUZhKbSCXN/F9uJuf0QCCJVyTyVQu9Gg5iXHNBv+5rmMND6Z00wu5NOqHImtcoEqlJIjeIUlP1//dN56esug6EvmobSbYb1z3B4IeJo5JdRaGsRn6c0/NFaphQPRusQLjfWC1cXaFvzgAwypjDmLldpLlRM2ipZ+sBGWTbfg59Qb+Crn+s5jmbX2pTITeQ1WHRkSWZyrEl0bDbCYU8KTVOiATztqyur0DxiMimCH5OoXeAsKgtrmpEojCWkyfUMlRNybGofnLuKUugnFSZNaxv7z2ZcLyFzU+G2WivwOvwOMD0AXrvKiwlWkhDD2cXnvrg09TTXc5RHa3PnxOXff4r3vx+/D3Kke1lUkRktv0D1tb8/ZCYuI8XX8TyT+Y1yTj7sV06qFVQhcpg7foDquGXM24l5VhBhqmaE/4WpyyhOVdBBVAdRvTJRnXzYWEvFV2Z6oyzXlda/poYmgBDdp0QZw8Ziw7qK9alQvi8usOtAQB7skVMsXoIXXduuVO8eWqkc2RmKDzhZStgsZUysuqR74M2rNjUwHKR1kNYrkta77nr1uWx2/leR7Wv/jbzOdinfgpH1xXkNJpAswqo0Jiw8pQkb5KyV1+oYtnOCfmTig6gOonpdojre5+0ZEgoqYapJPmxfr4+NjhPYc/jrnLXEnfK5ME1ZRHse5SuoSlpZEKTNk2IiPOcGcCXTcNtSSZ/jx2D2dsDzNQfJXhKXPBcO8bXl2HPvq9PBAIvvUW4c4j/SWhbyCHmnt/3UFDiGsx2yUX4eDP9qFX99w9+YA4Fd7YmT7cUV24urhgmyufJh5zHX2ds99yUsHLdBIdkldXdmn0vfIuK+l2LrDyqlsDtu73QcRBh2DspC/jayIuNjLjcMO5EpuTiptPSdz8Da2ED53cnHvMvXEIHFcb6J6itfg7nAngo4Qp+MSYkivaywG5As1l3yJXM0Ovd07weVxoaIowrlXjTMuI74d7iBQnL9+dNiMcZ26pvVyyW/xo7MziPLQk8bc5Wv+W4mRwPi60dw4pbmLIQ0pYr5h8NPHa52tjv7ssvVTx+HkMd/gcyOFw== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List tool groups with optional provider. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/list-tools-with-optional-tool-group.api.mdx b/versioned_docs/version-v0.2.23/api/list-tools-with-optional-tool-group.api.mdx new file mode 100644 index 0000000..2ee5134 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/list-tools-with-optional-tool-group.api.mdx @@ -0,0 +1,68 @@ +--- +id: list-tools-with-optional-tool-group +title: "List tools with optional tool group." +description: "List tools with optional tool group." +sidebar_label: "List tools with optional tool group." +hide_title: true +hide_table_of_contents: true +api: eJztWW1v2zYQ/iuEvrQFHCcNVrTNt2zpugDtYiQphiENDEo6W2woUiMpO4bh/747UrJpyXkbNgwBki+WRB55L89zx2OWiQFbaWXBJkfL5PDggH5ysJkRlRNaJUfJMfsirLvUWtrzZvIwGSSZVg6Uo/m8qqTIOM3f/2FJaJnYrICS05NbVIDL6PQHZA4FK6MrME6ELXPu4lncGL7AScJBaR+WFjmqICYCTDTXOiPUNFn5yTOcYsZopa5NBmOR3z/xjvHwoTswSEDVZXJ0lZQ6B4nvthAgc3yYobbajPMUn8lEC6S8zTQJjie1yrx3B0kKKitKbm7w2aGPm5/x1Oi6CvaWlUuuvcMturudlcOE13LrQxy1S9SU6QlrLR8wLud8Ydkrmv7KG4UPfpudVncXPD2h5VwBjOSYF8RXYcN7ClKrKb3Q0luiD638W11ytWeA5zyVwKJB2nFecLfZNtcYeAoYN7wEB8b+c/AoXOFh7X7HWa3l6123VBjfhY7dEdlaiL2G4XQ4YEFmwASSagrmzX/lxM3GrScN/FULAzECUvQ0cBWDzJkaulv+UQAuaQIINgvjS7smm2gT4ibUTIcMESxrll0mWsHZBCm03lzVUuKcnjLRJyRd6mPQZ2oHC9GHBgSr664Zr8/8E5dv2EnQi824rCEov+UzMWFKO9ZkizxZoS08z0WQH0XomnBpIXbuVYBbDzbb2kQSqKcTTpLqlHtHa+h19R9tQgoTobwyXnfuXT/sQSmhfE6o2JAoIjLPMqichwaO8U5+XlNpt9n/Rzx75kUBPV5ryVpjGE91vckoT4thVG+2i0ZTIjpJtRvcKGt1wtuL6nGIhiPiZlxhfvUcukFSpQvGp6iHvSe0JGyfZpuPdKRWr+r3dGwHGJ0FOEJPTRF0MlYBVVyRFj/tOllcYgxIA0CBObdkIJciZ4jdkkuEcAn5v3fSsI67Oi4LTa7tp9LLyxELs9GyHHytDE55KAt/MoaSBk3GgstsoY1jti6xwC/a3A9+zrwQWUG5kow2gisXOKuacb+PDy96Vj525zCbtqZqjCmhuLcsrNWhnQQeL7jKHmFlTDD27fxLY0wD09oiRp3GyKI0zICV2hB4KZ4+chEDbQUZ0iljOstqY/AsBB29noLgJsJttNbOi0Dt3XSH89qjMJvgmcvrcDw6HbJzXU8LucDwSKnnlp3/+gt7/+HgvWcf3PKyCsho8YVQ3+z3M8/ZeYD4Rp9HIr9hzuHH3czJpEBWsALFLT0g31BYLdqFaVUEwlTMAH9LXatAS+HL0AupXkj1nEh1+HGrYLKvhPSGWTam1p+6Rk6gh+A2A8jJbUQ27gArUynckI3w/IEKOUN1FMsWkzho2kr17q5KZcHMMPjoTqIS0PEWY1MruEV/O3yDxg0v1Hqh1jOi1ru4Xp0Ssum4fBHQvtm/oddxH/KtM/IhO6sRCcBLX5VSwMKD/RKdl/OWXlHr1znw3rPwC6leSPW8SHWwa7dHUMizhKDGp5asoQbsM7WTvmPc0ev5LovNhSuYbgISXdANt5tOasvDrVe3TxW0IBZTs+jfW6EPdl/+6dDrBRUwssP4/qKNTY+ZbW9PNzGoV6FxbjL196MVdwW+7M/e7ocGFsV9Hgqq14bcVzhXHe3vY/3fK7RFF+5JyUu+h67PboaZLv0djwV0qHALL3jSJp2raxojGHqF2k6X5NkFybOLBoy8adlp8+CH2dsdriHu2i0U85hX8cq+lW/uDOx3xTp/mCjAH9BB5ZXGtGQxD/jDizCYgUwAZk49NuUsM+EZ2LAoN/CdGnCpjWddf/GUGg2JPDR82mr1le6swx1ChX7EFLG5FH0stjpXlOsk/Vj5BhIObt1+JfEsRtr4MC8bMFwFvwc4YOwo5PR1uUzxEPfNyNWKPgfoUqxzYSnLbQB4p45PA/ZOVW9g0eeSvz3Ezx6HPq+nBLMruqkqMAnTPytQ0SB77K/bIqleddviyedPl8icvwFd6QF7 +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +List tools with optional tool group. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/llama-stack-specification.info.mdx b/versioned_docs/version-v0.2.23/api/llama-stack-specification.info.mdx new file mode 100644 index 0000000..e979bcc --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/llama-stack-specification.info.mdx @@ -0,0 +1,42 @@ +--- +id: llama-stack-specification +title: "Llama Stack Specification" +description: "This is the specification of the Llama Stack that provides" +sidebar_label: Introduction +sidebar_position: 0 +hide_title: true +custom_edit_url: null +--- + +import ApiLogo from "@theme/ApiLogo"; +import Heading from "@theme/Heading"; +import SchemaTabs from "@theme/SchemaTabs"; +import TabItem from "@theme/TabItem"; +import Export from "@theme/ApiExplorer/Export"; + + + + + + + + + + + + + +This is the specification of the Llama Stack that provides + a set of endpoints and their corresponding interfaces that are +tailored to + best leverage Llama Models. diff --git a/versioned_docs/version-v0.2.23/api/log-an-event.api.mdx b/versioned_docs/version-v0.2.23/api/log-an-event.api.mdx new file mode 100644 index 0000000..832fda5 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/log-an-event.api.mdx @@ -0,0 +1,68 @@ +--- +id: log-an-event +title: "Log an event." +description: "Log an event." +sidebar_label: "Log an event." +hide_title: true +hide_table_of_contents: true +api: eJztW9tu2zgQ/RVCT7tA7DhBg2z9lqbZbdHmAtt52E2DgJZomy1FaknKiRH433eGlCzash2nl0WDqigaRyI5wzNnDsmp+RhpZjIlDTNR9zE67HTwR8JMrHlmuZJRN7r8EM33olfrXg0mjGj2b86MJffUEC6nVPCEKE1SKkZKpyyJ9qJYScukxf40ywSPKfbf/2xwkMfIxBOWUvxkZxmDYdXwM4stdMy0ypi23HtnLLW5CdpxGHbMNDRcduvdYHBFfGsSq4Sh/5ZbwYK+xmoux7WuZ1qD867xHqHETJS2xORpSvWMqBGxMGPm2txPeDwh3E9acyotgQkTKov3zs4cx7eUi10t+9ZoWigJcyOTPKWypRlN6FAwEvRZcgctcQlzlvEOs/zt0n2i4ndyQq57H4vJxOD8kJHcsIRYBZGF3mzKSKo0g2liPF3kCB2q3DrjJmMxH/GYqDjOtWZgfsUvcIwmCff2roKAjqgwbC9C/nANPOnelBEuo7UA73YRPw/TBvBKLpORVqnz4eTqfZv0VD6eiBmERwh1b0jvz1Ny/EfnuI2gsQeaZp4ZJb+A6pW9NzQhPU/xyp8dmT+f4+xfHb5enzmx4JAVZALdDX6wSkFnOSsHxlGBCGM+ZfAzVTm0QWx5ypqkapLqhSXV4evK3gCYfo5MLzLLhKn1t8ohJwAh9hAzliBsmGzUMiJ4ym2bXAlGwSEL7KFjCkki4KVuF/l2tGmlMkxPIfgAJ6YSA3iQWLlkD4C3hd9YAUOTWk1qvaDUOgrXq/fIbPCc9D3bK/tFep3UKV+CkbTJZQ5MYDR1q9KQwcIjlQXA4F2RXgkb0VzYeoptG7hJqiapXlZSddZZ2yGFXJYg1ejY4GwGTLAUMJyh/8v+flRjxxTY3tk2Ep1qCk2Zxo63HhRYHN+oZIaefacEcubWL5DuFQZdqDE6pCS7HIEvTw1pNY3ZHU+e5ty15DAnwhMwhKqiXbpgkNwY8AlyybsxZEh+A+5ghExG5TdawCE2G8BdNcQ+zdaZ8KSHJwlsNFrFBnwFv7I/pBGTnvvOTEUMyAELIw5zy8y6OK3PkMd6GAq/5ns1EQweyTwdLj8ZKgUbJ7ncSIhofjvfog0f2KwFZxvANKNco5ZKSAMJDpDKYQK8BXGyNFAFTzSE1tlyBDYIYg4/dB7bHFC5A6pFwaKy/uVSlnuSwphhkA1zxL2+6A9616eD697Z27uPl39Fe7VYMgAGE3ONodoDTFweLykPWh8Uqg4RN3S8g9RicsGIpOgAK+yDQ8YARprb2ab9qn9LBHwQpaIG42yZHfQcKuNpOszxDUo3/LinWmK7cqsLBkBJltQVdKlfOvYs5V7oQJWvYWIV3la4BQAE5q+DwIArDvEaok6Cq3YOFZ9vIT+J4ajrDu8StHbI/0bQGkH7dkErZCKUseDRzuJ1fjbovT/90ZLlXu+kWBJ2I6Xq+H5AMYQpheN3XjDAIbk2qFtCWNsKOXN5ytCGj82y3cJiWgQgl9zuNgdsiWMFAyzypxjbz+AHCV3BAm+jcDyIzblrsEHiSgfX6FoZAe98I2mNpH1nSduyQ/uW/dn/uTvL6EwommzYW1Vbh6IdcdCURFzsJ55zDFrFEPMB+KbtMn7Lj0PPrgpXNqF3dXJx1x+c9AZbkFsa3v3CZBKC1A93VyVYqPVPJ+675TqEWyB8iyFyDYFDQHx9wKWxS2jNMjiVw2yMj4uGj3c7q0XA883C4Qd15vYIH3nj8JeSeMJF4l48U+O9hDtgQvRgoD6iW8RqYwxdHcjP30WjPOfvINXreIRBrLGofPhcDp1dvP3uDKoKc/VsG3HUn6IYV6zsDhkuE1fHAPKYPI5hh47/gzaiXMDoW1xUXxZnmHps0I2vCXUxg5UBz2TyjFADUotA4y6HY/sUpm/BVcCmiPTswuXbYp9AswwniGWcCvwnSzmN4DSCs4vgVFH5Oko12vPitGc+3wBZOd4POe+UW65N8dp00HmqkIPDL2/UvlZea/vKpzOiOTg1B6emuN0Ut19Kcbte42uUrVG2psrdVLlffpV7lQmNxjUa15S9m7J3U4X6+atQTenp1yg9NWXvRnB+CsFpyt6/nPb8AmXvcIpni32sFXeGQddkl6sRyA+chvuqN59WX9YvUXwGRuVuMfRgudzmnCyvUDr3q/4AAPPn9YmCX6NMuWuWGbUT+G1/erBvy+/P7ztLxtXV8GaLcVvTXOMX8ifWZt39fSpnrQmMwJKWEDSlLYhT/KUdq9StSuBe7guS0PFtmdY3t/gOv0Xfq75vf1ZdCFh8X746SlbJsRDX6lFwWosOO4evWp3j1uHrwcFR9+ige/hHu3N88E+0ct5anETWHQAWhdjA7KK0uqiFrpKgMy+qouh5GQwEhfQRFNIv7my4NQTP7oCoZ8f0YA1hvMqHlz384lMwJxzZTqglIK9TUELzSZKVP9TJIvSDvM0UUBQWD+nu+HEN5Nf+/kaC/Ef+6hGgbvygsPZ8wuQQSrvLKfXBh3gfFyvKGivR3qtzlTBh3IKA9IKVFTHxC3HtDsbSvIM7F7WGRbyw+LqfCUhYV7HR7o6Pp++NR3JBYJRPT2HgG9IUmzw+Dqlh11rM5/gYCKhn/vaHu080xLjd4CF4ApsBSGWk7heGgT/1rrUGXoOKClX9aghuvH2PE1D7zG5texvk4tVlH/c/w+L6SYp3qLqRpvfwEP/tRnhIdFA50XHPHiNB5Tj3fPVj4p//AJs6pbw= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Log an event. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/models.tag.mdx b/versioned_docs/version-v0.2.23/api/models.tag.mdx new file mode 100644 index 0000000..355cfc0 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/models.tag.mdx @@ -0,0 +1,19 @@ +--- +id: models +title: "Models" +description: "Models" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/post-training-coming-soon.tag.mdx b/versioned_docs/version-v0.2.23/api/post-training-coming-soon.tag.mdx new file mode 100644 index 0000000..ff77c13 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/post-training-coming-soon.tag.mdx @@ -0,0 +1,19 @@ +--- +id: post-training-coming-soon +title: "PostTraining (Coming Soon)" +description: "PostTraining (Coming Soon)" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/prompts.tag.mdx b/versioned_docs/version-v0.2.23/api/prompts.tag.mdx new file mode 100644 index 0000000..668884f --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/prompts.tag.mdx @@ -0,0 +1,19 @@ +--- +id: prompts +title: "Protocol for prompt management operations." +description: "Protocol for prompt management operations." +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/providers.tag.mdx b/versioned_docs/version-v0.2.23/api/providers.tag.mdx new file mode 100644 index 0000000..d4b4183 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/providers.tag.mdx @@ -0,0 +1,19 @@ +--- +id: providers +title: "Providers API for inspecting, listing, and modifying providers and their configurations." +description: "Providers API for inspecting, listing, and modifying providers and their configurations." +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/query-chunks-from-a-vector-database.api.mdx b/versioned_docs/version-v0.2.23/api/query-chunks-from-a-vector-database.api.mdx new file mode 100644 index 0000000..051ae99 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/query-chunks-from-a-vector-database.api.mdx @@ -0,0 +1,68 @@ +--- +id: query-chunks-from-a-vector-database +title: "Query chunks from a vector database." +description: "Query chunks from a vector database." +sidebar_label: "Query chunks from a vector database." +hide_title: true +hide_table_of_contents: true +api: eJztXG1v3DYS/iuEvtwVWK/ToL22vk9uksMZSGKf4xxwSAyXK3FXrCVSJand7C32v3de9Lov3nVTFHGgfEhsiZwZPpx5ZjQksoqc8oU1XvnobBU9f/YM/0mUj50ugrYmOovOxX9K5ZYv0tLc++tq+DgaRbE1QZmAM2RRZDqWOOP0V4/TVpGPU5VL/CksCwWC7ORXFQeYWDhbKBc0K41JcGecdE4uYZgOKvdHzG/NsEZdTqOzD80UH5w2s2g92vFujzh+uymBVutBSaRzOVPwe6Kmssz6T7qwvdT4W66NDNYJlCbsVIRUicpggesbi/NsIZdefGQxHyMwtpJ4cOWly3Zt1/vr17UqEiRAfyKDpBfa0IupdbkMOAzfnNG40xWqW/9zIr36x3ejFb5Zj8VbGxTMgdE4P7ZllohUzpXIlJmFVGSwyODRIQ5Zq3fh2rf+BmxDNfxaFFabgD8ES2aDt9rSxSpaA0oySTROk9lVR9FUZl6NwK9/K7VTCew4ab4F83TIUDXIR5BxeXs2GrfnlYltgk/OIgZky1Z+LBQOVEmFNQEtvag975CdfZEXJAOmS7EhvQIEtlIaQSt4DAK0xtqvOlCQvhe84Atwx601nlfL6voshdPnBVFQn0IvhpoHnxVCKIUjiOQd9jYYVcv6Q4CSmg6eKPBhOENHZYUmzE+6K0W7KziXb2WOcit1OfAsueTqWIoY2Gxgs4HNDrDZJlsMpPb5pIaKm32BFy5TEGVJV04njWyVfEO1NvDbwG9DtfblEdtQrQ1sNrDZUK19VaR2TLW2vn1ozM5oqpXUkGGnbyQWqY5TEYMvTYAlWklk2Ig33Y/Q3SzMYtQhxsHEXAW5ETrN7u/GbUclacos6ybFibWg33QfmTKfKNd90mkk9krW7eS6vl1vQvGmMhuCy9tYywBrXWigrwYU5riFzjLEpPQY0kyWOURgxkDCriUlRaI2U+UgOhWhosDYJKkyzYEearOyDRMvC8ZONMKQplsDx+JiKowNAoJsroEUYJ9ag2ObFyWuKoO1OTLKQ7yo5I4m3+nkOPZlLC5eMhwamQyl1FjMAd8qnyBDjcV7BArtnMj4Xhn4uTQxL0SHJdnBBnQdZ8/e9Jbb2Y+3lzebe7J3NwQu4pe+yl9wHXXc7jP0UP56HIwAYDfimr3zKvS2baaMcuSNCGciLC8vsXGZY9iCHAm2VlFMcNbv/oAtjdiQAiIM80Rl1sw8pFh2Gk6wRwnmsRtsPBK+BG6hJAaJaiSmOlOikCHt8Am4LyT2qa48NXYKMbgLOlc+yLzo6Ed2mkG8bJGoEbaOmGYeOEJChyDgE4tUmY47LcCiSg/pLIvkL9GZSR9EpawTDguYZRfH4cxjH6JwCg0okmbOlgW4ekYuxcc68HymEPWO9mAhBPT/lTvOgGZ4o4iR7Dr4S87iGGo3+p5mdBQ2nHZHbHqc2pYImYL3Kf+bb4fu1JnAXhmv+VDswB6j4mZ8jXlrSUWAfWYmnez+DO0d1MkmHKmNEwKpwrm+x3EbqbuXg/9EXbXITWWHqqG6GqGjyZrIH1dExU310qSI2w3BOyooDjCwtkaJ0kVT0XhQSNkC3EVuZa3t5Pta+9CVVoWOU6F0BjOcszkB8xsexBJRxpAXjzkz3Zfvr1Wm5hJylmBRoNvxUXBSffIoCRHemEAmPRJaPtdtjO3guuNAeQvl+gWvnlaOllXYgOfsQpYK2e92nWLf0BcciAGokRm1mUPuTTAr5DLDT1Kw+k870wZuDqU/Iiz+fXNzJXi0wI+tqC3ED5HUK+eQB3DwCMDwqXWQ4ss8l27ZUAeNYbbWvGinJbjYlD/m+D3pIRcJUh9Bj6yZR6NqzOEQ12mZS3MC9JjICWTdzpyeOdRkgG8w9L7Duv5eV6bfCO4u7Ew94KdOqzmWyw4jkFsMqFlObBlIuS9UDFk/FjaOS0cF24Zdj/Huaofr3WrA63g5wbQHPNfzbrTh/OpiLK5tOUuzJWxPltmFF9f/eiF++PHZD1znf4KEz55R+xe4eqvvZ5mIa3bx1p4jPb+KnOc/7Y6cONNITCk2F4jvrIXJZlkLrgJypudQgsgccwJhC6lsCKohqJ5YUD3/qdNOAU9/g55eRZbvhtb/bMntSPUpVgpbZdQoxPqQOpNjcZUpyEwigPfImYQgqb+QKd6+35epoIKYw+ZjA67EBglIhr0pjfoEeGNtoSoYhtAaQusJhdb33XxFTUT8mn3H3t7qr8LrfNvlazCSsbgswROUzCkrTRQkHmPpsz6pw6vp7W6dUzwgeAiqIaieVlA926XtiBCiKEFXkzOPq/kvfU5dXKL5fXPpe63+9CJjd3x8gftLJ+EzWjkUd8tQQcr82SbLjYu6nxNWrPgumRzdg2wbfvVebxiP3kNfmAQxf2UP14mHI93hSHe4oPLkznKHCyoDmw1sNlxQ+apIbbhOPPDbwG9fBr8NxDZUawObDWz2dbDZUK09mevE1JjD+PFKujjFyKeGHfUdd/5PAV/wpWBcUNswrd2g7kI+Zud6DdG6fbnvtg1fD6AdamUEVyq+2pVa+DUqLF0hwHuT8Nvp/NtT1nGi7SmLH0V8ROkJRGLvKA2hODs9lWZ5koIAlZxkmczliQ8yvh/HlrOdVzGQWFjSxJd1uHy4xXfYCr5um8av2l53v+nben/Vq20fNL6wphOGqSW3qJB4jfaId2iPeFedBFATGibiYnhn5t/u8D48OPG9IwTZPdToSqYEU92T9h+N2Pgj8R4uzlMmoaTg6ZYtSNFu4xoWXY6fylh5Fiqd+miws0+3ooPdFj7BWx6ZgsUg1bFVb/D+Il+gx43NJbXbDRcXR7b2e3B0WvnHzq/CAbnitMikNnQLlpM+u9kHxr1xtK4nozvhiNUK5b132XqNj6vNx6MGOtKa4CZ/wFBLlUzwfim42L1C/6ho5eSGWW4us5LidfMcAkOXZ5zHsSrCg2NvOyFzdfnuBgZPqrMOvDMKT51cwEP8+yyCH/j+LjEQPVtFmTSzkiqkiGXin98BBgNZow== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Query chunks from a vector database. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/query-metrics.api.mdx b/versioned_docs/version-v0.2.23/api/query-metrics.api.mdx new file mode 100644 index 0000000..68e85d1 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/query-metrics.api.mdx @@ -0,0 +1,68 @@ +--- +id: query-metrics +title: "Query metrics." +description: "Query metrics." +sidebar_label: "Query metrics." +hide_title: true +hide_table_of_contents: true +api: eJztWF1PGzkU/SveeWolSGi1VbdIfaAfq60EKgvhYdUi5Mw4Gbcee2p7gAjR377n2p7MhAkQupVWleAhTDK+3+dc2/cqs8LVRjvhst2r7PnODv0rhMutrL00OtvN9tjfjbCLA+GtzN1RWj/KtrLcaC+0JxFe10rmnETGXxzJXWUuL0XF6ckvagFNZvpF5B6CtTW1sF5GqwX3/VXcWr7AIulF5e6XroJfvXUO3/Uc61bDmJSCaV4JZmbM4znJXW9lik+Fcj/uAWl9uP1glcyfc9VsKB+W3lAADbwoJK3i6rDn2IwrJ7ZQ4G+NtKLIdj9FT1uLp1uZl16RyVjb/aDwpt29aIhx50wuuRcFu5C+ZDxlcEQxrIrsS+fJy5jYgaQvpUvCzAlLzrZp+A9V8LISzvOq7i2VwOdc2EFMJ1pesqUAuyiF7mEipfmCO2ZFbmyB5K2pk26q6Rrdoc5NJTpFK4Bj3McEdP5Cd6OlH0LgQbXt9LWuJrWDOr8D3w4NcrOm1g52lWBESVbTGib1stTB51SyO8reCTuSzktrtFFmjv6gGKXTPiyyxNQlUZdYGUR2HNE0DKvnOHmYwuk5ekc4K0hlFfd5iSSFkn6jxsgg4/GSPyyq0PV6EaxrsoNI2heMWi+XmhwZ1iYEhoiuyaHf1/V0Aik5A8QEnEuNlMoC1UGAamZsBSd/WoMHLH3jNuDlX5PJIYurEWEhqCgpP/e1x/fWwvmweAuAdaWxnrmmqjgqlBgowpqLUuYlkzFolA0YR8CM6/Q+2AloQIbVppbjajKtjEZsrGwqrret4AWfEqM6mRV3yJLUiFnnG0T55GMdwfWU7bGTo/0UTA7np4I1Di3WG1QW0uIcLcdYgTCpnqFyjE9N44NxV4tczoAbk+eNtQLmb/j1EDCnCrfVWiavh++QpluS1x5C2MyaKviwd/hhxI5MMy/VAuVRylw4dvTnW/byj52XgaziEr0uIqPFF6De2XvDC3YUId75syHyE3Oev1rPnFxJsIKVEHf04I2BsF60il1smnN5jn2FV6bRoZEQQx9J9UiqX4xUz1919iZA+gEhPTHL9an1j2nACWRIXOZCFJQ2IhsOfkzJSvoRO1SCwyEP9PA5ti+cELFztjvVi9t2Kuxq5yg+0klUEkgPAavR4hL5pmOlSGl4pNYjtX4har3o71cfCNnwnB1HtHf2E732hpBvk1GM2McGSBC8CrvSVGDj0cYjYXiX6FWIGW+UX3PBv0PxI6keSfVrkWpnnbUNKBRYQlDjc0fRTIQSdLVakP+r/oarWrp3ORqE1dxyfBWWJNuBULq3nqWhiyTJmvtyw8kU1TDcL8lAl2ZvG6R9QLrlxOA0Lsbe/MYUC1rx8/hr/Vk4w97P4bBrk0C8lt4WGFVQFw9RiuX3qpxbrhsFVvvFZmO1nsBdesPTWVQ3VCt0UxFuoGse60189kP0kEUSJlNxgAArSDQRdLScR56FQUNA1P8wlySXggM/NKBsheMhD/cvKCHHuAcd70jdazz/Rh+vv9PT9/W5y00FvklHPTQpJYsNdY5wi4sDmqy340Hzj89Je76vH5kexEqtGTulGga/ZlLRlGg5sMEuvUjJjuOsNSOoybIeLRooUmJxAE0H1NGDO3XL5RVY3zqOitfo0B9XWxF+gAelKai5mXDVDj1uNxufPxv7toeOU8zjq15TvMbieLSPbbOx1KNL7+vd8RiXjO0SCkWxrRSv+DZ8zr+OUPyMepwT6NqB4BB819b50ym9o8521PXA9yt7xLKH7fR7z86NrtFhs0/6JbdvMrTr+Uu5RJruh44AhMbTsOPPTJxcx5TvU5zsmOJkx2lnDr2a1MFOxMT5szXMo4OMW9nSef+Q0dfsS+4ZGsW5hJLPmt34w6lJhGkFkpNmuFyHm5y0oJ6Nu3RBOKY+bWc8Fy4q5VZ8pqmkMjYcQYbKpzR1UTiUWD5vvTrAEU9F8BOAcF7qetdwp10JvLe3DVemknlx6ce1Qh8Kc3YbznIRop9iLpcgzbbSlk1nmN3+7g1QERZJ5OpqiivsiVXX1/RzgAcBjyqOc+SUKgk4FNLRc7Gk362OPzlKhHrKNjgIrI2q3R70ogc8PH4VixvHkIC7EsdRtCtyM654G53ZnpCeTsPgtHC91Urs5bmo/Z1rT3ud4fDj8QSLp+lEUtGpnth0QUcbfAZvTUhO2LnCb1egmZ43QAreR5309y/D+gg3 +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Query metrics. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/query-spans.api.mdx b/versioned_docs/version-v0.2.23/api/query-spans.api.mdx new file mode 100644 index 0000000..e62e3bc --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/query-spans.api.mdx @@ -0,0 +1,68 @@ +--- +id: query-spans +title: "Query spans." +description: "Query spans." +sidebar_label: "Query spans." +hide_title: true +hide_table_of_contents: true +api: eJztWFtv2zYU/iuEnjYgtxYrtuYt7TqsWItkifswpIFBS8cWW4pUSCqJYeS/7zukZMu3xFkLDAUSBLEjkef6feeQZ5Y58rU1nnx2PMteHh3xR0E+d6oOyprsODsRfzfkphe1NP68XX2Q7WW5NYFM4A2yrrXKJW84/OJ51yzzeUmV5G9hWhPk2NEXygM21s7W5IJKOgsZ+qukc3KKRSpQ5R/f7WHVUBW9hT44ZSZYuOzFJ6OuGxKqgMlqrMiJsXUilCRYRHa/lwUnc/pGWVEGvikfxYoRaWsmXgTLGmrpsGO4s80/ncZvUv8stqtMQqO6PaHGSTl+pchLpYu5e0ZW9LjKP5tKmn1HspAjTYI3ibRihOVRIYc/prrnp6MaQIIdnlX5IF0YBrVZIQyvJGDDmaf9uGrVigEeQkhVi9uSzIrWEU2SR2SKb1HSi+6D+sbKKF9SEaOb26rWFKhgA2SAulETaCNQZVGopOBsCbLW0Ok4O76c7zCN1hA3/39kraboYm9JNSLXf9J62nuSmNN70Fpyf3X/gOt/0XT/RuqGgaScF8xqCY+R7IUDoqIgmahCjmwTerS53+bnWGpPeygv141yCNfx5ZyrPaq1sFyCzBXeq6DZA645a3k7WYFcNFV4fOh+2m4V4GnwJio7yNaC8EH5IOw4SvMCaAFfWohfc8ETWBvIKfk0L2M96/mwXjzXPOpeLAVf6L59sP+ezfhlU4UewGQ2ARgWtxLsN8ioKgQqRCU1U4GK71ewkarQ9BGvIHYCdK7VksHgTKTV8KygWGRTVB4rQ++c4/LGi/c4uaV1KHFNVUnkxY5jkiiuuS1VXnK9Y6eRLJRCrowASHof9cTcI7J6V81pNavm+o1qWy7Xxd6eJXNYkzLw2eQ7eNmj4Yn4dP6hdSaPjUM0ngo0DmQWu+mGRGUdOoBJpY0198lIOdpCLmyeNw4NAV1o2a4nETVluMvWPHg9VMcwbQled6QQY2eraMPJ2fsDcW6bSamnSI/W9taL8z/eil9/O/o1UpPuJBfWPr4A9YW+N7IQ5wniC3t2RH7LnJevNzMn14obaIntXE4QcovNZtoJZqkAwkTdoDXIyjYm0rLtKc+keibVj0Sql68X+gZA+kdGesss36fWP7YBJxAhusuJCg4bkw3HKXSmSoUDcYZTiucz71TICdqW0Hjpuk71alun8uRukHyEk6lECA8DqzF0h3jjZNUG95laz9T6oaj1qt+v3jOy+eR8kdC+0N/S62Qd8l0wigNx2gAJJKvYlUaExmNsvPcVHb0KGstGhw3X9QcEP5PqmVQ/FqmONmnbgUKRJQw1OfHszYA04RLrpmz/sr3xgtbdsuKQBldSsJf3XaWYoDe+scWUDftO/JmPDoZjpZOy/zoC+0rTx1HGjXeuU2ALgyqpFjZOU2y95XhsK4REeQZ5vFwzn6xg59nIVbVkmooDTtd8uWfcTNhkVKrVO/FbaxIUT2vWH4cQ/898ZFOw0kwEjib/KR1xcPt4Eok4NzG0nX9bg7BhzJF372IxS8ni0UCaToBgqP9+w2BjOdctvuY54y/d/CbtXoyxhsEOUWkaZ3aA4zzCD6n3bfGCSL5GLSuu5N2woDqUO/QRlor1qmoqEfd0pS044unOU7KyTr4tQdgyxknX0FhfFmKDa4h9olBa/JvVNl5Va8nuZYc3Lw5DV4MOYwh46hUPBz6CvXFc08oQ6uPDQxzK90sIoGJfa1nJfdTD/OsBsBjh6glVToVp3Ph7dxK4vOJ3XInOFzXr3aKmbqg5l235WPCXywCTl0VtBsZlt/hqKYNHseuNbcxlG7QPbLu4YNvFRdudZIt2djyl9ubFhmynIXa/rcl+o+1LDqUMAjXxRkHIZyNWfnByoHhjJ1PUFvjyOBjE24xy4JhLnapgZjH43FjmDFsWCt5/5omcti624XXhI548aDRmJyedVR9xzNEJ4QwCnBk4Jmn+vtptltzudZfVdS07At2Fw1qjErH0CJlZC7HLFMc5yBheEWZIE0OJV8xmI9zYPjl9f8+PYyFJXS4em0actEseF5c444CEC3y8TYbtD9iOebFeb4FcXtOOkzwHMB5ce9Wjy9npxQCLR22brfioeJw5eYuH/Pc4i+Dk3bEIxWezTEszaRB6vE8y+edfq5VqwQ== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Query spans. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/query-the-rag-system-for-context-typically-invoked-by-the-agent.api.mdx b/versioned_docs/version-v0.2.23/api/query-the-rag-system-for-context-typically-invoked-by-the-agent.api.mdx new file mode 100644 index 0000000..f43d1a7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/query-the-rag-system-for-context-typically-invoked-by-the-agent.api.mdx @@ -0,0 +1,68 @@ +--- +id: query-the-rag-system-for-context-typically-invoked-by-the-agent +title: "Query the RAG system for context; typically invoked by the agent." +description: "Query the RAG system for context; typically invoked by the agent." +sidebar_label: "Query the RAG system for context; typically invoked by the agent." +hide_title: true +hide_table_of_contents: true +api: eJztG+1u3DbyVQj9aQLYaydo0ov7y81HGyBu0o2L4hAbBleiV0wkcktSthcLA/cQ94T3JDczJLWUdtffwJ0D+Y9XIjnfMxwOR4vMCDvTygqb7S2y57u7+K8QNjdy5qRW2V423v/1j0aY+VjYpnIs18pxqaSaMlcKZoQzUpyJggaEcoyrgtXC8YI7nm1l4TXC5bNZJXOOcHe+WgS+yGxeiprjLzefCUCnJ19F7mDhzOiZME560hIwWomPp9nel3aJBRLUNLvcWjO2AZwf7UMgai0gyWTNpwKeC3HKgevOm1Q4byQ+1VJxpw1DaEyfkmCiOKQT9YjtV+d8btmRB3OUAbEB4rWcN6ZaVco++3P8IaIiQAzwo8hpQCoaONWm5g6n4cgezdtZILrLnyfcipc/bi1w5HLEftdOwBqYjetz3VQFK/mZYJVQU1eyCph0dgTUXUetXCfXLvWHQBui8cNspqVyZFA62JTVjclFdglS4kUhcRmvPiWITnllxRZY79+NNKIAjRPmYyBPugpRA3wUMlnhekWjet6qXBf4Zi/zAlmh1b9mAieClXtZk6C5ZdHyrqOzC/I9wYDlnPWgB4GAKrlixMFtJEA8RrtKREH4XnuG34M5rvC4H9hKbZbc6X5O5MSF6/hQ++JeLoRQvAcRvOutDWZFWHcSKKFJ5IkArxanS1AGacL6IuUU6Q7inP/Oa4Qb0NUQJ8kkFzcNEUM0G6LZEM2uiWb9aDEEtfsHNUTc6gUGTCXAy4oUTrKNcGP4HFUK7+3aTG6Ib0N8G+LbkK39zwPbkK0N0WyIZkO29l0FtZtka5fHV81ZwfPk48yz8ZQdrq0GnhpdkyT/xhoiSqytDa5R7nqxrEkUVVNV6Z430RoIVekr1dQTYdI3SZ2wk5Gu7p2Xx5crEm1Ja6ubjE9045bcYfBAK7uVgltxJILvllxXhB4qsWCinMHUgDwpzF5dlB2hJQCJP66r9notAkTr2DmECanOeCULDAs1rzCoA9kPVtW1jrsmceQMgrGYgtb6LP92ePiJ+dkMw1W2NOXrfO+tMejUOHkLBGZLbRyzTV1zEFpwc0FzzkuZl0x6po3kaL4+HPpxwkOGAYKuborZz0bUlVbAGyubmqttI0AZkwri+XJNhxzapiGKcZXfgMvEEf3+7JnJgfiJYI0FW4AdLhoGq7WBvVv5TRoxL03ZzkQuT2XOdJ43xsD2IHp03ca8g4ajtlrhJdZOYtogvHg9sYwj+5/ej9hYN9OymoN6qkqfWzZ+95r99I/dn0YoNHHB65m3jGhfYOpLfL/wgo29iS/puaHlB895/mq95+SVRG8rcXvGH05rWKzmETBCBUOYyjMB/2vdKHJjJ2sxONXgVI/MqZ6/ShISsPQDtPTgWTZ1rX/qxif04iIXApNNSrU5pP2U24/YJ9i9gSAH1sOnsIuxCgZN3KlebNqprDBnoHxMYRtMVgAy6KZR4gLk7eBJBDEMrjW41iNyrRfpfkVpOOadn721L/EH99pfNfkojGLEPjZgCYLXtCtNBGw8SjsQGIwF92pPRysn/SsAD041ONXjcqrdddhu4ELkJWhqfGqRG9jqqjFsN5i0HfcppnMb0YgnMzu3cKomAyJnuXA/owGBp1RAOFiY/gboJn4Bn4IzYZVrxg2Hw5owiO7YixK21F90Me+1wgwdNUNVc6hqPp6q5lDOHO5ohmg2RLPvI5oNdzRDR80Q34b49r3GtyGwDdnaEM2GaPZ9RLMhW3sUHTWHbTdJRAV+ZAU3eUl1xBAMpCrEBdhYofOmhlkWRXcGCtXmpJicyCItqK+kif0yX4+GD9JSnPHwyD/QtNn7Nzah5ly6UmLLTUb0ngC9p3K6Gt2SqvRrmtIYX25e1jmJs2UfDeqPplwboDzmqVC4AFjfREMXcRdbuxrj4UMl0NE/Uk9J33VUHtyjR84W4947fggLf0BZW4FiCxvoqo9EXGwFy2cfjdr17X3AVwjYATfoAgzkLu60JCsx9zeenNjO9GvkzGtjhcL1Sgo8JZ1OicLun6RVVd0RXHy+lYJgESmnhrh/gzshzHdiPKy4mjYYrmkt6qOxtNd3seFCCo71DK/GbxIg/cy4CyFEn0G4+O1uHwEDydWzu4VTz3lCYGIGHz4c3M8EAMA2BqBigxHcKbdMbj0HV38Urk6+eW91DQ7/+B0+zbs2YULt8IsTp78JZU+kOgk3sBsaDYI5/Lj76mWfxgN+IeumZr6jmPRIQGMuFgCPIsq8bNS3TQ0NAc+L65F4MOmNPWGg1ydX20W07dApvKBs8fJIhaRzjy0IyiikmDByEJqDYSj2CcPbzUbWsy7BISFM2o4Rel88bP+My4raHQBGLkpdFZD77UXy2JNnQel+uTZwloOscatPLXviJ8QE2TOO81rS45S2VbuQuXs6YiFM7cGJoyebK4SzQTpHyh9Y0BM6G0rIw1fDss+bcT7JL8iLV//5178FJNOg9qOw9ijbgt/fxPwcpIAPGsfK+cRIeGz5SOYnh/nWEAQYE7psS08ACL88qF7TuSfwILTWGK6+CXPTXNrPjuEMVO8xhMNCSzHZ83j8Lsx/yJTbmNOO8cfn/ukqnkY9BW1MP8IFsZYz47k7OeV5d/MNnxQkSF7ursPg1zO/niSEHNtcI8Uj9pucoq5B8Y2w1AzsW3HOBQzQea+kGdtEYRE+K7AjdtDAuWwi2NQIbI/ECo9iu3csK6QspmYwfjf2il/97CCXoJGcVwxnsHeNRQN4AiueRvXnqW08RHruZULtZkvNdl7eWL1xldcxr2Ylv1K3u6OVIP2X11BQ7ES4c+yn26XPK56N4EctOOwLWlVzcoPgb6R6YbfYs/6EcMKO48EkpGphQ7QE2BPtyjBpdCd1e24TNf8VhLFB13F4rV59ZTHX9UQqoDbwgDLo8nvXzBwd995p3hAN/g+jwdJz763fITA86sDQyeA9qluRv6Hwtynn7yTmKzn0mk/v4iHiNjTlbUm3W4hNwIfP+vzHRySEJQBnGuG/jyw1PGYzTR8ozbgr4Wnn7NmO07raNr71dMfw6Ta+2PFfVuK5HNvDLWVSdM2Ulc7N9nZ2uJpvlwBMFNtVxWu+bR3PMb/113JW5I2Rbk4L30Qj/3KMY9hUOl62n75ddtW27aNL1+yVn7/EkePVCvHmum3wumUBJamCZOyKU50/uaXnrxerh6WHOBH5fH+ZVC/z5EC732p628bLXVI3tlvT1HhURoWwz6gQ9jk0XcfSN2rTe93ZszURDXvUbadbm6f94ylk8kuIr2cSgBwp1vvjkKVTwR9CCl3fWXJagCLRw41vwMabNYZHWQM8CeuBciOO8LvTShsqMa0Cn+AHdRUcCw0WPTxVeMSoLO0IaOU1p85l5ROCh2ih7sgqNdaHAB7UjNN3wK78HUi42PXe+sVrLPVXMhTvsVnwB3QM9Eucv1jg0fdPU11etv7iu7/pK4QJGssX/Cq5FLxAawNfhaDqj2PI2/ahD+i0U+CdT781HHNxv2I/z8XMXTn3OIlDnz5+PoTJk9B+Hszf8HNi6Rx+4yGOZE07N71bZLHOBeMeJv79F9Z55hM= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Query the RAG system for context; typically invoked by the agent. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/query-traces.api.mdx b/versioned_docs/version-v0.2.23/api/query-traces.api.mdx new file mode 100644 index 0000000..a12cafa --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/query-traces.api.mdx @@ -0,0 +1,68 @@ +--- +id: query-traces +title: "Query traces." +description: "Query traces." +sidebar_label: "Query traces." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETrtAEqfFFt3mlna72AItkk3cwyINAkoaW2wpUiEpJ4aR/943pGTLsdO4214KJIfWkobz+d4MyUXmyDfWePLZ0SJ7fnjI/5XkC6eaoKzJjrJj8W9Lbj52siB/1okfZHtZYU0gE3iFbBqtCskrRp89L1tkvqiolvwrzBuCIpt/piJgYeNsQy6oZLSUYSglnZNzCKlAtX98dWC3rlQ5kPTBKTOF5HocH426bkmoEj6riSInJtaJUJGIOrK7vcxZG658I80PKmQ9gvXgSeJXkC5QiQflV8bi26ugatpmCqpqidxyemg/St23P8ZLKKkbcVORWUUicppKwybIlD9i4LeT+Evq38WDtibKKF9RuSfURBS2bjQh1OwO1mVZqrT+dFCyidSekGq6bpWD5NHFqob3KrCWo8u9LKigOYYIxQ1vjzuPHDUANdfETKOjvVeCbqloWVo0MlTCToQU7AciE7Jw1ntRtzooSAt2OOLZH3Am1029V1iB5dEgFslQVL21ayaLgHAgp+T3JSJSYRDoFuJthN1/EExHiWrADyn0mocI4Y4d+WMbv8cM2C4LN9ILZWZSq1IAy7XUDBMqfx7bUdHQDnmtoHZKbiOuf8bjU5GkEVoZKdPl5TFivnWOicjCe8iFr6wDCdu6lqgMJwURU5S5qVRRCZWCRrlMiBwGcdP3aCeWH6nVu1pO0mxaW4PYRNXW0uw7kqXMga3BmjV32JIyiNkUO0Q5YOex+Hj2vgumgPM5idZzw7GoLFbTjERtHXqVSbRnyzK3bYjGfUMFGlghbFG0zhHM3/Pre0DcVbiv1jJ5A1zHND2QvH4giYmzdfTh+PTdgTiz7bTSc5RHa3vjxdnfb8TLPw9fRnbSrWSKD/EFqK/svZalOEsQX/mzI/I75jx/tZ05hVZghaiwnHsOUm6x2Mx7xawVQJiqGdqmrG1rEi9Tv30i1ROpfiVSPX81mMJA+gdGescsP6TWf7YFJ2Y8cwuiMm5+QDZsNTCaahUOxKkm6XkXMRdyirklND66flK9eGhSeXIzFB/pZCoR0sPAag3dIt+8yaIuDU/UeqLWL0StF8N59Y6RDc/FeUL7yn5Hr+NNyPfJKA/ESQskkKzjVMoJg8fYeEIpe3qVNJHY6W457H1D8ROpnkj1a5HqcJu1HSgUWcJQk1PP0YxJU40cztn/dX/jEW15zALQpZMQJccLL1NSMBxf23LOnv0kAsmAguZtoKuJ0snY/71A+ULzx2HGk3dpU2AJoyqZFjZeNdjmgf0xDt8glWeUxyM1E8oKDp6dvG+WTFtzxukaD4aBM2WX0aruH4vfWJOweNKwfezc21h7a+hkAhXLkEyrNSSWz7m12HqY4SsYzckN33QODd6ktA5edPm824AEhx3d4UBT/JT2ODh+fBeLuDYxtX18DyZhy21I0X+L3SwViy8H0gUFGIYBsO1yY73WHb6WNeMfy9uftDxu6HZo5qw4yg6uTlJzap2Jmuxk4mlXVUm471Hb9bmS3FU+34Edy4Jvs8RqRD7/hq3Hqrr1RicdSGOnWVU9uJbwAj2ksnjMGhsPrXxjhafR7Nko9N1olBzhe7K4T/AR9q3j9laF0ByNRtif71fQQOW+1rKW+2iNxZcDoDIC1xMangrzuPCvflNwccnfuCedrbrX21V73dJ9LrpGsmIyNwSmMavqEHK4qvDhsDgX/arLOPsmNtaly9h7dlucs9vivJtRsoM8x5zKNHu2BSM80v3acJPDcTvUHO9p0RhnCko+GXHvD/uHBDUyZWOBSo/tQTzTKAeiuTSvSqYXQ9ZNEkBYKcj/iS/mtHVxGG8qz/n+QWM8OzntvfqAzY5O9GIAYOfAOTGyXoJoMHPW4h7MmA3BDuuBbsOo0WhIrD/iZdEB7CJlcgkx/p1AhjIykFhkschxdPvo9N0dv44NJU27uH/KuW4XXMsKmx2wd4WON8m1/TE7smzam6OQ22xacVwU1IRvyl4O2HJ6cj6GcN6N25r3jEeZkzd8t4x/j7IIzXixywLx3SLT0kxbZB/fk07++woxv8sn +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Query traces. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-benchmark.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-benchmark.api.mdx new file mode 100644 index 0000000..245da3d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-benchmark.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-benchmark +title: "Register a benchmark." +description: "Register a benchmark." +sidebar_label: "Register a benchmark." +hide_title: true +hide_table_of_contents: true +api: eJztWN9v2zYQ/lcIPW1AYqdBi65+S9oOK9YigeM+DKkRnKizxYYiNZJyYhj+33tHSZacOF627iWA/RDLJO/3991RWSUOfWmNR5+MVsnpyQl/ZeilU2VQ1iSj5OLPZH2UvN61NclROPy7Qh/EHXihzAK0yoR1ogA9s67ALDlKpDUBTWB5KEutJLD88LtnJavEyxwL4KewLJHU2vQ7ykCCpbMluqBq73yAUPneOUVq5+jo4LZbf0wml6I+LaTNkP0PKmjsyfrglJk/Ev3oHDkfDx8JED63LghfFQW4pbAzEShijGfuciVzoeqgnQITBAUswDT70c6a9QdQ+rmW69NsWltDsYm8KsAcO4QMUo2iJ7PlDltShmI28hlR/nIRn0D/Ks7E1/HnJhhJzqcoKo+ZCJYqS9K4QFFYhxQm1zNWTkBqqxCN+xKlmikprJSVc0jmH/hFjkGWqdreZa+gM9AejxLGj3KEk9F1W+G2WpvkTTf1q9P0RPJaLIuZs0X04ezy00CMbTXP9ZLKo7W982L8+3vx9reTtwNOGt5DUdbIaPFFUO/snUMmxjXEO3+eifz1mqN/ffpuN3OkVsQKkZO454dgLQmbZauYtRIQ5mqB9F3Yis5wblWBB1IdSPXCSHX6rrM3IaR/YaQ3zPJ9av1lK+IEZQjvJWLGaWOyQUChVaHCQFxqBHIoEHpgDkQSTZtu0PDtzVOTyqNbUPEpnUwlpPQwsCqD95TvQL+wScOBWgdqvSBqvenPq0+MbPJcXNVo7+w39Dp7DPk2GdlAXFSEBIQiTqUUafAYGyhhtNfQK8MZVDo8ptg+xQdSHUj1skh1ssvaMygUWcJQg7nnaM4pGzkB59ZzANsOj3GuPNGVCp22xwaMeHBQIG2whmmdHZqS5zZbsov/E5M2Jm9U9s9o4QH66UNb1o1sDZE6jJjKDALN5vAfdDaSrLHi8jF9+qaiei8tq7qZVUayln4nAOdgSUZUwMI/tr7eZb7RJzb69punDC5Uhu7mZ5LXKtnO4jOM/oSpvQYIa8DZ34WZ3bRdJdbgxYzguZEwldakbPM7tZauaaa/ZKoiRddf2ZTmQRF7C40n6+nO+rWu7wvw3zSfrbpuoXkX9notqeXyhu/tS2PsBp2J4CqsU55b+pmUNr5YlhBy+jVcvAJd5jBEep0cpl3vIOtxnvuY8spxG8pDKEfDId2jj3PSgtmx1lDAMbUweTuQtqCksRw1JhWWUfBDO7yvp7zHLWPcNZePXRvcxncPbD16d6s7WHnd7k6fZE0nv4XwbrmHy3UcbTMbIdqk/DNHK644WnHVjKDYCkmUU1WjZPFqB0d4Yvut2QX9adrXHHIILY/8NyMefOh6gPG1HE1WWrqMeJr+8ZVFObp3uHocZdxj+KbiZiDR10rB4TfDI8W6OGsfK0/53wuapq+DeevVF7rLaF83Bqo6XQw4J4YGxp6JshV/b4I8KdCQL+B9GJaa3rHYXoTdqsHqddJglU4zWukr3Zp1jEk+tlqlBJmvTq/XvExoc8t6rsUrU8q1vGZ253S/ofbAOL1FOpK8r908nrAzfFxXsUU8HHrcLWqJMymxDHvPTnvku7y4mrDfzWAt+Jo4Shzc0SL/HSX0YMvNoIlrq0SDmVdUEdqvdfLnB5uYcrc= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a benchmark. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-model.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-model.api.mdx new file mode 100644 index 0000000..4425407 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-model +title: "Register a model." +description: "Register a model." +sidebar_label: "Register a model." +hide_title: true +hide_table_of_contents: true +api: eJztWEtv2zgQ/iuELm0Bx06DLbrNzX0sNkCCBI5zWKSBQUljiw1FqiRlxzD83ztDSpbkOFkXCBbIwjnEEjXvmW+G5CoyYAutLNjodBWdHB/TTwo2MaJwQqvoNBqyC52C7Ee9KNHKgXJEw4tCioQTzeCHJcJVZJMMck5PblkAsur4ByQOGQujCzBOBDUiRSFiKsC0aK0zQs2Qtqv9RomfJbCGhU21YS4TlqHpujQJflRMSp5zZh1P7qO11zdHFjOpaSYifTFdLgNWK+go20fH2Vempx0R+MId0wtlu6pIdJD1WCaoMo9Ob6OcMoPvNhMgU3yYY7y1maQxKeaOW6Dw20QT42RaqsTb0YtiUEmWc3OPz05rWf1MZkaXRchYXrjozifdYso3ulKY8lJ2V9oejtG3TbjI8B7jcsGXlr3xDG98VP3jhs6Sszk4TjbvKiCepoLkc3nVKSWt4HKKkdhwqFJKFLZ5j9Ep4Kq9hLGLfea249pa4cbwZXuhsmR9t972d6iWrDGP1V40tRPCRA7Sw6TOaRNHKfOdUSRKqpYQq7fn5xcMZQJaj+rULKy/i3pPlkcQvGGgbDrhJNF6TI+Jb71+KrxTLi30sEX8LIWBlCS2oNst/MqIVhY7/m5rfuTvcKsi8KHAZ9KGnnLFhmcbipmwDtAgAuO5B/41Ab+PrpAzf+zqYqEqEd7WsQW3yDrnUqQU0JxLzFUO6ct1OGxErrQtOoFiZz5oXbP+Ho+vWKBmCbrnQR8C9W+d5JsxVGFEjAhjNtPGMVvmCOpl3WPA0ywykWRMBKeN4Mr54sSghu9ej69rx4XcV3OgJtVSK/SNZWXO1ZEBTH8sgbV4OuaQJoEthatkDy/fXhahMN+xIbsZnVfOJGh8DKy0WAVOY2aRG+aAJWKoR1M+feYYj3XpvHJbQIKVmzCdJKUx2P9gy67fAUKV4Tpbm+C1Ct2H6Yng1WOXTbHPehuGV2d9NtLlLJNLTI+UemHZ6K8v7OOfxx/7FDR44HkRKqOuLyz1Rt9nnrJRKPHGnj0rv0LOyafdyEmkQFSwDNkJkhhyjczY+CrBJBULYSbmgL+5LpGGYityOIDqAKpXBqqTT42+MVb6BVV6hSzbhtY/ukRMYITgIQFIKWwENu6ASZEL12dXuPlAgxxWD59x2qjiR1NPqg9PTSoLZo7Jx3ASlPysw9yUCh4w3g7foArDAVoHaL0iaH1oz6szqmzaMl+Ham/0V/AaPi75Ohhpn12WWAnAcz+VYsDBo7TfmqY1vDY77EdH2mcEH0B1ANXrAtXxLm17QMijhEqNzyx54w9mlozvGjuqDlyY5Ly+CSq44XjSA0OcdyEqOB0/63RJpr0QgsIBcp9rFRqarXubKqXhwOhLI7jQ79zXvIz4rRuh/m9fCe0W3pH3v74a6e+6G3n+LqST0v/y/mNTMi1RNT68yPoA5pHVcDpTQshjpvE1KrQ/pBXcZfg2mL8f5AF9vSjs/azPXGkIxJlzxelggLvQowz5ID3y151H/rqzn+gcY098CGvhlp7xaz36bu/oGwFv1ED0W9NEGgw04dsBkB0fu8utCt3Kps+Cb/dT7eu3Clvr5oZdV22ZV7eTFICQ+/n7HXihKWY7/Zy3J0xbsr9crWy23xXb+sORCf6oCiotNA5oixPRb+OFwVlsQov2t200vc2UJ2CDUG7gu6I2q42fP4+Fx3TkljiRDJ/VVoUmG5oE5hKHJcVEYTN9otN2fG911p3EVcQdPLhBIfG8QXp8Ea2qWrsNEc03vZ4qipZXqxhPKjdGrte0jLVilqG3++1CTDm7JYhnONvp5h6r7B6QJPoSTDoahwvAOZel7xPbjZ9aRuAYJgkU7lnauxZYri6vx3RlXQ0Xsh1XDV/gIv0/jfBB+xD5zufXVpHkalZi5PF7kEl/vwA7HreT +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-new-dataset.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-new-dataset.api.mdx new file mode 100644 index 0000000..94f28ad --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-new-dataset.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-new-dataset +title: "Register a new dataset." +description: "Register a new dataset." +sidebar_label: "Register a new dataset." +hide_title: true +hide_table_of_contents: true +api: eJztWm2P27gR/iusvqQFvHYuuOv1XBTB3u0essWlWezuoSjWxoKWaIsXilRIyo5h+L/3GVKy5dc4L2gTwAsklsghOTPPzHA41CKxwpVGO+GS/iJ58fw5/WTCpVaWXhqd9JNLdsU9d8J3k06SGu2F9kTFy1LJlBNV7w9HpIvEpbkoOD35eSkw2Iz+EKnHwNKaUlgv40IywyRyLIVt0TpvpZ4ky0A8BYl9Anemsql4ktlxwgP9sWG7o5MIXRVJ/zEpTCYU3l0uhcrwMAW3xj5lIzxnUWzqTg0NfBpXOg1a6SQjodO84PYtnr0xqv55mlhTlVHeovTJMKjMQWGt6TIx5pXabmvr/AH8MjNmjfwdxtWMzx17Vo94xsbGsvrFBVVUtjTuqLTo9xfecqnR3iuEc3wCPEAw5ar3rhKOlr/g2s0ATN3ckDXNw21eb+PCxK7PRcMTkzoL1qEnTHqHV9hNJjJWgUewG+XatTayNRY7GTQ3lpPKBhsLArcWAH9GizdjCPYhaztkBQ0ylZUbqNA7WKSfPaO2kGqJnHLNRlDEyEPFEHUMG2Cc/X5302XX3UmXXbBBkntfun6vV8xnYuSkF93UFHijSbrkR2qQBELlxpFup4ca+qmb/n2EZf/6fWcRf59q51wOkmQJ/nmWSeKSq9uWOsZcOdGB37+rpBUZ2UWQMMoLdL30igQG2wTGfQRqW+7LldQ+50dFhyo/FyBrZm4DodAAEcPveiC3ls9BCK0Wbl8U2q+Sxa4l6UqpNuMj+Lbgut0EpxrBH1otq7izxdCuApZDAuiwJUnHHOIQFCk1IyFXBvTIFoOk8clB0mePeLdgjp4HCbzLDpIOnmpjiM2vhFKmw2bGquxPAzDENkZx56TznKg/OHS4ZMNPMq+AVcu+7vB+koFtaQIsgFYSbSE1RychWBvU/F+8oMnrNQtsUgQJCDa9+Rwpvs5Ise3R54BxDhifHzBo9WaW1gxoLYTnNOwrx35H1hVrrJFgJz/6KJW3EvLNrLrToNHkl6vUraW84aZ692W0dfsqow3sElCUIHKdMZ6mMFJ6azJUBgpKQauY/4XgRlhCru/3HVTIH0gmpLFsxinlxGCZ0TQFV1ivENmXO8LA/H3VjhmU4U6C/jbZevXwcMsiNTLaLJhdra0PbRnX1hKoRIwTAHO5sbDuqsC5Y97k2yLQzHKZ5hQFSGgr4ZlBwYi1sT+sE8wIYVedunKkpqWV0ZCN5VXB9YUVgH2kYGrrMRvs0EpSU4BIT5Dyz2/KaKN/YZfYB36rhak3CgSojHkDZDFaTAUr4NwQk/CMhsFHpvJhcVeKFEacMpOmlbU4ooktvj7GJ2qEG7RWymtZe1DTAeU1J+u4wxEPl7fY3u9MNcnVHPAgTM4cu/v1F/bj357/GHY/8Z4XZbSMxr5g6uv1fuYZu4smvubnRMuvPefFT/s9J1USXsFyDHf0gNMsBut5MzHNCkOYyKnAb2Eq0JBuZSHOTnV2qm/MqV78tF7vAZb+miy99izXdq3/mAo+AQ2J96kQVMCgRS33gimcP3yX3WKvB0Me1sMn2LmYQqdtdqofDu1UyLumAB/qJFcSlLAAm0qL99C3x5uo1XB2rbNrfUOu9UN7v7ohy6Ys9T5a+3r92r0ud02+UUbWZW8qWILgRdiVRgIbjzYhS80a91qd8Haq1kcmPjvV2am+Lad6vm+1E1woeAmZGp84kuaquTDYqeLfiQlO9ECXMy1mzSGSrnxKbjlOesLSDMOoHeySP5tsTix+IU/6v11f0F5c7r3CQPzR1NYPpbT9q1PJY6OwB22AxEGNDQnaVFVoNpM+R87gQtq86iQH3Zi6yxZsq2DD/scVmw4bsmUQep9mj4jcUDYih3M9/dC4pnG8cayP4jYDIz//pjohwhhhkfJSemwg0NmvlkLJyyjAmpdBcotgB021ed6C/bNhOlmSTwaumDMNPyO5/2lyza6M6H4SjKvR7JWZhQB6w3KhSjZHKutNxucvd+c9wBsh8cw1vMVxw231NwsSAseu9hoAmuu9bW+71q5CVA9VYr9FG8MIKSc1RQl9084ToPJH/Pc6RlJHHkzgUGSp5bUyihse+qdW25cnTLS/Gn/KyA9X60+aJq8mEwSSMU8F+FCKFxwmk77tOUnaeMdfOsRp/48QcfbOSWXTOGl8Cqb8dUQl/Dvf/H479znni5zzRc755vccKc43v+eAcb75/QI3vwT9oStfAj1gD8RaoyI4r+cNGZAJDle/7v12ct+yN1c7CfbNmCpyrL4wzjp0RgLZTCpFrj4RWlCpPJTrPsIMdu6a27ZQFyrqUkZzGxaQXs/hbSUiyLnBazi9h1qGz/HWm37XW307iUVCcdIFYCtLNRUKwIirXM8vcowU2UVIoy9CHk3BOGxFTqSIh34eBl41EedxSH0Ui+/WlZLrdU1nVek4XNBYn6BqVOLuEzab1tetLRPegrNFREWyMFOtv99IEHZPgrD7ulTG649aSQsR7+l3e0yAKotuo8bG21W/9swh4Ndm4Qaabf3hXCfCAVvorDRS+3jCxiySTtc2ls2y8PEolZDpLOPipNyKAR3eVfB0b3YnH9E1qBIQBuqsuXpNH/q6YPak9YKH4pSOKcPh2teGBlq1riNDasS8eO97pQK2MU0I5dZofo+1dlu1ODIy6liEbfp3q5ZLaob52HmsuoWC7ogQfKSgkAue0cfTMLy3AiTJL5G1i4cYQakqESLLdkmOgkwccZmmovRHaYctD7p9c/8A4lFd9qMvp2l35TOK1/i/n9C5LKgqxMrQtkgU15MKOKA/zkl//wVusQXB +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a new dataset. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-scoring-function.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-scoring-function.api.mdx new file mode 100644 index 0000000..a0a6a8c --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-scoring-function.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-scoring-function +title: "Register a scoring function." +description: "Register a scoring function." +sidebar_label: "Register a scoring function." +hide_title: true +hide_table_of_contents: true +api: eJztW21v2zYQ/iuEPm1A4nTFhm35lqYblq5tgiT9MLSBQUtnm61EaiTlxAj833dH6t12I1fu0KTah9WheC987jnyKIr3gQaTKmnABMf3wfNnz+ifCEyoRWqFksFxcP53sDoIft706HoOTMO/GRjLbrlhQi54LCKmNEt4PFU6gSg4CEIlLUhL8jxNYxFykj/6aEjJfWDCOSScftllCqhWTT5CaFEw1SoFbYX3zlhuM1PrJ1DtDDR2bLr11/X1BfO9WagiIP+tsDHUZI3VQs7WRP/QGp13nQ8YZ2autGUmSxKul0xNmcURg+tzOxfhnAk/aC24tAwHzLjMnzs7K9JvuYi7Wva9yXSsJI6NzbOEy0MNPOKTGFhNpuEOWRISxyzDDqP84dz94vGP7IS9u3ydDyZE5yfAMgMRswoji9KwAJYoDThMiqeLHOMTlVln3KQQiqkImQrDTGtA8y2/0DEeRcLbu6gFdMpjAwcB8Udo5Mnx+yLCRbRK8G7K+HmYtoBXcJlNtUqcDycXZyN2qbLZPF5ieOJY3Rp2+ecp+/W3Z7+OCDS440nqmVHwC6le2XvBI3bpKV7505H5qxWN/ufnv2/OnDAWmBVsjuKGflilUFguC8WkFYkwEwvAfxOVYR/CViQwJNWQVI8sqZ7/Xtm7Rqa/IabnmWXqqfWPyjAnECG4CwEigo2SjVtgsUiEHbGLGDg6ZJE9fMYxSWJ8qEd5vv2ybaUyoBcYfISTUgkQHiJWJuEO8bb4F+QwDKk1pNYjSq1f6uvVGTEbPWdXnu2V/Ty9TtYpX4ARjdh5hkwAnrhVaQK48EhlETB8lqdXBFOexXY9xT6neEiqIakeV1I922StQwq5LCGq8Zmh0VyFitD/M5MhuWloGE23L2EmDCYtMc13ZtO894jozzVPAJ+TuhsPFS6ZL1S0JH/3lVbe8ngqxyJ6mDu0nJ69LILcdtvzxo9q5OlfE+6iewPD17FZERg203LsFd4HSsL5FGF6aLRF/7YjCCaW+Q3P8slum7MvBf2VCMltnu4jdhLf8qVhH3KZD8Fu/HdO1ch+5bRcU2vb+kXBDWfZTT3eJsPdCLKEMOqLhcySST69FljUmjpj4WX6YvHWaemKBdoEjdPS/sCYKIW1p2ygUW/rDEcu1BePF15NV0Byq3sEhGvNlw04qpbOYDiRvlCckJKuQDiLe4ShFKtwqDV1BsLL9EXi3GnpCoW3uUcs3NJTR6Js6IwDSfRF4RXq6IrBq6vzt3tEIJOiBUHV0hkDJ9IXhHekpCsKzuIeYQjn3I5DRZUVGR0LmWbNHNneozNMG1X0he0UlZ6WOs9IZVcQyR9W+cOcP3tB87NA9sNw3/B9IXRfATU+w+J47IrEddQ2Puy+arWkey9gpO8a1e0EmvOCkRclaLTDqDtO+OSoLd+iPCqyXnuCWwbCy+17dP5rKJ2LCbAodHuD8qRq6LLg7Q3Lk6um8+K3NzBPpKouht8bj6dUXgfFq5nvvcrOi+LeSDyRantLMdwbnqEKr9eT65Vyf4C/y8J8vXjuv+p9h7X6qjqV8hIbdfmvTZygU/a5V+KI+UJEoMe9XukXWja+28+MH8/DPvQw3MkOIWa+5M1/wbk4TsbcjD9m0QwafFt70Ha5CMQaQNWBzQHjnoAtZW08AItsYlmrm4YZ3I1RnXEbhQk3ImzsbvKjJemIYxxzqL4g8XGiIuhw4HgWISXpfFcXAXj9+g1zwkUEuGGcOaV+l+St5mFOUkwrwOmB292OHE8zY1XCvApWqCjD7e35QZRjItMwdrDUo1kWxwLVmHUv2gdQdOJGGmiAcGc1D/0xJpfmFnFw54aYqUCfnETleaKf82Zo3Z2wjafliV5nR6pI8wVqd0G+BTGbo51x1ZRAJNwOKEQHZoh2yOOx+2YF2zidM/Jw2ZyNSreKc8bN0wg2GopzbRwlb009ujRqzAOzfnpXN1aTRSjp+HFJP4p09baAh3Om1e0XTLhNLm9mwbag1OBBRp+YVyTbypjtc7YHAwUPuTn0XFzLc5xCpmKGwSimo13nnlZ+V3PP2oM+c09L2dede0iA1pw95eiQod98hrYj3iEfXWgvHLV2TEhnhHlWfoWE9CxvvI0qW/qkYKHl6+beQP3/mfoPEv0FRW5HirtoP8ztL3rH3+BZ5+3aUCoPpfKwEH/zs9GjKpWbCbvzVDRUzkPl/OgT9luunHPa7pyYQwU9pMAjqaDrb94fskB0TZtWNn737syDDOcJ158YLHh8QB2xMsxvVShkgRYRVpL+QCBNXa/RjjckGm/2m842v4lvzh3ug/zWhYTibqvDozJidQYr4qqdK/wzSJW7/5pyO8e/jhY/HeVOHFZxOwj8DTvj3sNnmq5KzK1Nj4+OuFwezlEFRIdxzBN+aCwPP41ClbitjAFkvbBLJ/iymDve39AzOqW+rO48/FG7qtE84NhWSZfNrcsCa8votpOTSkPjXKPWXB5D5Dpb01Fjm9HQ1twTVI82FvHvi+c3W6eqauq5WbnLQFPl3CrKO8KeXRH27Cq/tMNz2lDgPGCLnzYkAN1xMo3bPn6OKLZCNc2WDl9zqMwHyVr/cWbAXWQGGaVKSIubJukueQo6atS+KIkosehul57ykDYgpJRr+CDpEo7S7nbSuvIJXciOwYOQe/WGgPdzHdE48Z8uSb9LfuD2TevmSnnb5iG5nAgWa7AjjK6Q7oMH7baZPonee5zX0wijS6lCPe7vcTKDdzperagZk0Av/S0gd9tsQkHFLEEB4BEV0pg+nwC7BKfe0cN8oXCfGdAK1r4iRC/JvMRJGEJqP9v3pjYhXJxfXdPSml9DInJT/c1vKdHw/8cB/lBpuYS6tvsg5nKWET/zD1bcban/AGY4o1E= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a scoring function. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-shield.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-shield.api.mdx new file mode 100644 index 0000000..9f745cf --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-shield.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-shield +title: "Register a shield." +description: "Register a shield." +sidebar_label: "Register a shield." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETruAY6fBFt3m5ma72AItEtjuYZEGAUWNLDYUqfLDiWH4v+8MKdmy42RToJcAziGW+Dnz5r0ZkavMgmuMduCy81V2dnpKPwU4YWXjpdHZeTZm00qCKobZIBNGe9CeBvGmUVJwGjT67mjkKnOigprTk182gHNN/h2Ex4mNNQ1YL9M+ssBFZCnB9sY6b6WeZ+s4eIFD7C0aZ4IVcCuL5wc+0Z8a9jsGGehQZ+fXWW0KUPjuooP4sEBrjb0tcnwuuOcOyHgnDE28LYMWEZRBloMWVc3tHT57Y1T7czu3JjTJ37rx2U2EzHnavNujgJIHtdfUB3xWAes8Z2T6gHF1z5eOtRPIc2557Q4hzYtC0kJcXe1gbjRcluj0ZoYOSuFSm/cc7Qeu+00IU44x6rVssN20cGv5st/QWrK+We879ttlkyz7nV0YXcp5sJE/LHoDHqxjpbHMIwCdq+unPCq5cjBA/v4I0kJB4ezRapcbLRMwGl56RUZOD0M/Zo6X4Jft9r0wVNwzwTXLgQUHBfOGIdnFHWslMURTydg/DkkoRfRHAOfZPXdM6gVXsmDoa80VulxD8evU5Tz3oc8NicvOIyi7Zv0zm12xNBrdKCBKJgH0WDO7Uz9aS4GiwchOhMtYz1yoURJLZsoYQohj7ispKiaT01Zy7WOMEcrUH/eJVPFcqpfunEbT1spo9I1Voeb6xAIveK6A9ebsmEM7SRQk1+IFXvYIO2ZfJ59bZ/Z4YAFnwwJYbSygmxTPxGuem+ATnRsQyEzBjBDBWswesGfXzxC9jXAXrQ14PYJHmJ4Ar8v5rMQsFW0YX30asokJ80otMTxKmXvHJn9fsHd/nr4bEmjwwOsmMaPjF1J9u98HXrBJovjWnhcyv1XO2fvDyhFKoipYhdMdPWCixcl62S1MqyIR5nIB+FubgGMIW1nDUVRHUb0yUZ293+43Q6Z/Iaa3ynJ9af1rAmoCEYIHAVAQbCQ27oEpWUs/ZFdYz9Egj+zhc44iUdhpu0r19qlK5cAuMPgIJ0kJEB4iVtDwgHh7fIMWhqO0jtJ6RdJ6269Xn4jZaDmbJrZv92/lNX5M+Q6MYsguAzIBeB2rUg5YeLSJn55FJ6/NZ/6j89QzCx9FdRTV6xLV6aHdXiChqBKiGp878iYdyBxZv2vtBObSoVYjwbp7iO2BEefeJFywPn4wxZKM+1UaivsdvFw4dGbfHj+7qLanyEiP5MVw59riV22ApZ3eunWH/3s38rLVd9d7jTcO5FfvcmEHteFPym0Tq57AOnIm9nbnn0js7VxvA2AD2lAZfM0aE89IDfcVvo0Wb0auJf8gS99eLqIWLImo8r45H43wK/CkwolQnCjFa36CAhR3Q2Fq9JvmoaykX8aJf3Wl5/qG+oj2k61APvZEvOXflhmH2Hmgd6+5Y8c65tbSRKK0KH0mg9mUDGbTNgfy9iKNvE2xWrw5QEsqGW4nefJ+Ou+vHK9oWvPcN832/jB9QDwXgi4ag9XQYfmJ38zSYuGzKR8W6A6jUmlLLsClRbmFb5pymrEx2T9ePKfzrcL0b/m8s+oLXS66pB0MHFYmwkQjF5/KajvO97LY4dEt+T08+FGj8POedoqcWbXcuk6Yum1qJQZR+2qV48ngq1XrNTUjN+wyZdJYnnMK2zXJqcJaSre0yKo7wCHZRTLqZEa703AVoib30yzJM80YCwGNf3bsTU8dV5fTGV2wtqmcrmix1fJ7bKT/5xk+mAhSzDKxbZUprucBwcf+tCb9/Qcti/93 +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a shield. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-tool-group.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-tool-group.api.mdx new file mode 100644 index 0000000..0bf75b3 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-tool-group.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-tool-group +title: "Register a tool group." +description: "Register a tool group." +sidebar_label: "Register a tool group." +hide_title: true +hide_table_of_contents: true +api: eJztV0tv20YQ/isLnlrAlhwjQRrdnEdbowlsyMqhcIxgRI7ETZa77D5kC4L+e2Z2SZGyZNcpejEgHWxyd+f9fTPLVWbR1UY7dNlolZ2enPC/Al1uZe2l0dkou/grWx9lL/dtTUoUFv8J6Ly4BSekXoCShTBWVKBmxlZYZEdZbrRH7Vke6lrJHFh++M2xklXm8hIr4Ce/rJHUmuk3zD0J1tbUaL1M3jkPPrjeOUlq52jp4LZbf04mlyKdFrkpkP330ivsyTpvpZ7viH6wlpyPh48ECFca64ULVQV2KcxMeIoY45nbUualkCloK0F7QQEL0M1+tLNm/R6keqrldJpNK6MpNlGGCvSxRShgqlD0ZLbcYUtSU8w6f0KUv1zEJ1C/ijPxefyxCSYn56cogsNCeEOVJWlcoKiMRQqT6xkrJ2Bqgo/GXY25nMlcmDwP1iKZv+cXOQZFIZO9y15BZ6AcHmWMH2kJJ6PrtsJttTbJu9nUL6XpgeS1WBYza6row9nl+UCMTZiXaknlUcrcOjH+/Z14/dvJ6wEnDe+gqhMyWnwR1Dt7b6EQ4wTxzp8nIn+95uhfnr7Zz5xcSWKFKEnc8YM3hoT1slXMWgkIc7lA+l+ZQGc4t7LCA6kOpHpmpDp909mbENI/MdIbZrk+tf42gThBGcK7HLHgtDHZwKNQspJ+IC4VAjnkCT0wByKJok07aPj26qFJ5dAuqPiUTqYSUnoYWEHjHeXb0xs2aThQ60CtZ0StV/15dc7IJs/FVUJ7Z7+h19ku5NtkFANxEQgJCFWcSlOkwaONp4TRXkOvAmcQlN+l2GOKD6Q6kOp5kepkn7UnUCiyhKEGc8fR0KhTf1gTascBbDs8xrl0RFcqNN39lJjzuQFDHixUSDus4ialh8bkW1Ms2cf/iUpsM5r8Kot/hwtP0PP3bV07fxNIUiAxmWRkIQu0/0FrK8o6A5eQKbRlLVqo8vor6qI2MiXh8TCDlU/zgzmQtkVUzQ/kSLx9oDPBEqt+CslsuYda0r9jNzHP4gwTbcge3jUdvClzDBns3O0Ldb8vq8xovJiRDxsJHZQiRZv3KaUUQfeXdKimaPsrTbJ6K2AtLPsLjSfrm/VuaIXMo2+pa1IIoaJ4HAdZg3Ntcreq+zP53QLwNvB6aW9ZtmFi+zkXedrp8zYggwt9aeg1q0385KvBl/Q2XLwYbuxxd0q3SRdzHCy3hdL7ejQc0r32uCRZLI6VggqOqaXk3we5qShLLEeNQvplFHzfDtPrG95jBo87rn/o2tI2VzsUb5GtW77PkMiBrpwtnNZxYMxMtNCk6yP7LK7YZ3HVNPbYX0gtB5yKu3ixh0I8B93WRID+jOpr9iX4lu7uixb3fjR0MX7stkE4mqnxQ0Ba4oVNTb5ghvL8tzPI0SWlYPGL5kZtbJxgu8qn/NGuaKZZmLdefaIbgnKpf1HtaNxyTjR14cf69FYCen35YYmGNZ5IPqwVfbuwxQifVYO065TbHtYIGIwn3lqtpvTl89mq9ZqXCSl2mUZEvH5MuYLXTMWS7grEZcbYd6Qj2bvk2/GEHeDjKkQ+358fTO0kcZbnWPtHz9706HJ5cTWhw9NmRlV85RplFm5pkf+OMnowMVexQ8W1VaZAzwPVgfaTTv79AKoiMwM= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a tool group. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/register-a-vector-database.api.mdx b/versioned_docs/version-v0.2.23/api/register-a-vector-database.api.mdx new file mode 100644 index 0000000..914a306 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/register-a-vector-database.api.mdx @@ -0,0 +1,68 @@ +--- +id: register-a-vector-database +title: "Register a vector database." +description: "Register a vector database." +sidebar_label: "Register a vector database." +hide_title: true +hide_table_of_contents: true +api: eJztWFtv2zYU/iuEXroBvqTBiq55S5oOK9AuQeIWGNLAoMRjiw1FqiTlxDDy33cOKVnyLXGL7qGA8xBLJM/9+3hELhILrjTagUtOFsnx0RH9CHCZlaWXRicnySn7DJk39vxskPSSzGgP2tMyXpZKZpyWDb86WrtIXJZDwenJz0tAaZN+RWkULK0pwXoZLUmBSuREgu2sdd5KPU0ew+IZLrFjdM9UNoOxFE8v3DEfB9YnegnoqkhObpLCCFD47nIJSuDDLMQ6Fik+C+65A3LeZYYEx5NKZyEtvSQFneUFt3f47I1R9c94ak1VxniL0ie3IWUOE7aqGia8Upuj3cSP0GdmJqzJQY9xdc/njr1YyrxgE2NZfGXkbsqplBg4FCkIQT7HELfkYNXaP7wI1nwObCnMgjDzhlUOuramoMGG0q8aE7IA7WQEQ21QImKmWOh1i+fN2k2z0UoIZBnrWKOHW4qMazhKkVKuLjsom3DloIcI/1ZJC4LK3YHdKnZqpGzmbXtwWFUvvSI3GnJshPd5tSrLMoY0Oh8AxbgW7FsFdt5G3abBDTA6iu+Pbbwc5aQThZ1n99wxqWdcScFQQ8EVGilA/DzCOs995fao6t+j0SWLq1mGKQwsjLl6DoLvrEXnw2LEOnO5sZ65qkCWzZcYCWvuc5nlTMagreTah6xyXc8HO4+k33O5B/ij5biaTCujMTaWVwXXfQtc8FQB68isuEOWJHKc62yPKH+7KCNWf2en7NPVhzqYDJ1PgXgmiG8WUBpmgAy0gGFSPUPlGE9N5YNxV0KGYM6YybLKWtyQYM2v7+FGXeGmWsvkdbAe0rQjeU0jYRPc+IIPp5fvB+zKVNNczbE8Spl7x67+este/3n0ehD2jQdelBEZDb4Q6q29My7YVYR468+eyK+Zc/xmO3MyJZEVLEdxRw+4d6OwnjeKSSsCYSpngL+FqXAN5Ra3gAOpDqT6xUh1/Ka1N0KkfySk18xyXWr9ayrkBGYIHjIAQWkjsnEPTMlC+gG7VEDdzCN6+JQjSRRO2qZTvdrVqRzYGRYf00lUAkwPAavS8ID59vgGdRoO1DpQ6xei1qtuv3pPyEbP2XVEe2u/ptfpJuSbZIgBu6gQCcCL0JVSwMajTfhaFQ29lseGjUPaE4oPpDqQ6tci1dE2a3tQKLCEoManjqJpzmaO/F/19wqm0iFbsc5rx2e65Si5xcMmTpOa25gibJVnRszJz59Ep/Zou+32YuM6ALPZnl+bGq85H1ETQxv80D3AaOf5f/DjR31SKnYf94OVwbP3OftlpFEx2Of2YJtK3bkJ2cBG18f/o374QbUWxPfQd8WjzeJ3WNzgv6ZI2hyzAn9ald5WgAPIhdzga1KacBQruc/xbTh7OYwm+yKlXSN+5RFpFkllia659+XJcIjfm/0cZUH0leIF7yPVs7tBZork8ZbkkMDSz4PgedPkbm5pjlh11fLvXbtdrOa/c7m3jvltUx34Hq3hrl2+Dp92ZgcIOlePtF0HVNQZ/0Bxs2uKm13XmzavLxMpaREls5dbgEM9zq3s9rzbf7qafc59gx73RbO1P2yoEA6yoEVpkK4uXEGhFmmxU9u4gQdSEpfthGfgolJu4YumTdjY0J02lad0IFfYryyfNl59pAq4yBusP7ZSykmdzac34ZUsdDbdZ8RqFnp48MNS4QmFbAcwLmrc3sQsd5CLMCN00tRiQYo+WfX4SMPhci42gfCRkVItEZgogF8EdH2NiL0DXJK8jQ72R/EmccZVRY5sdIjHXiNxmmVQ+ifX3nbId3lxPaKb57oLEbhx1PJ7HKT/Jwk+mJCw0F/C2CJRXE8rrAjOR5309x//dWs4 +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Register a vector database. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/rerank-a-list-of-documents-based-on-their-relevance-to-a-query.api.mdx b/versioned_docs/version-v0.2.23/api/rerank-a-list-of-documents-based-on-their-relevance-to-a-query.api.mdx new file mode 100644 index 0000000..7642585 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/rerank-a-list-of-documents-based-on-their-relevance-to-a-query.api.mdx @@ -0,0 +1,68 @@ +--- +id: rerank-a-list-of-documents-based-on-their-relevance-to-a-query +title: "Rerank a list of documents based on their relevance to a query." +description: "Rerank a list of documents based on their relevance to a query." +sidebar_label: "Rerank a list of documents based on their relevance to a query." +hide_title: true +hide_table_of_contents: true +api: eJztWd9PGzkQ/lesfblWCoFWV92VN0p7OiQQCOhJJ4KQszvJunjtre0NRIj//WbG3mRDAoQeL0jhgSS79vz8vrHHvssc+NoaDz7bvcs+7uzQRwE+d6oOyppsNzsFJ831aRonblQohTKFysELb12AQgynwoGGiTQ5CJ9bB+IdCQEcZsbv+1kvy60JYAKJl3WtVS5J/PYPTzruMp+XUEn6FqY1oFY7/AF5wIm1szW4oKKFhQzdUdI5OcVBKkDln5+NZsNtZ5hCm8bgcNyiy+clCOvUWBmpBU8SdiQCPi1s3lToBz7l38rUTRBa+ZDd97JZEK44CB1NpqmGjyh6GLmRsxXLrmwBWtgmoIq++EfqBiMuHSmdAMf9pgQjUjiHGgVYnCiDKNW4BBfl+ZgsGUCMHeCHm2vsZ/dotSwKRfZIfdKJ1khqD+TSz0Y5KLLdixS+ZTcve1lQQcMMLF8pSw993RMewaDJYRqDH77RIbor00McIFpE9imiiyIOMc6UiUUJMdm+tyYaX+YzA27JwZYNS07OaLLCrZ+Yv0AhJwN+X0W1CAceJ24kZW4itSoQi6KSemRdhUa9Gpd8kKHxa9Dh7/PzExFHixxBSYlJ8ZjN9cGhl0tTvzmHxvPgHobDl5gi4Zuqkm7acgp4zE2pciws0WmnJHIMHRbSpPeshxERpNLrao6jSbW2Bn0TZVNJs4VEKJgxnTkL5pAmZdBnxNDzut4d1xFM78We+H56mJzJ0fghiMYjKIPFzOJsmBCxmcSUT86ckENkOSv3NeRqpHJh87xxDgjCi3a9BLwpw222ZsHr4JnD9Ejw3AKYyYa9k4O+OLXNuNRTTI/W9saL07/2xR9/7vzBhIVbWdURGS2+EOpzfV9kIU4jxOf2rIn8xJyPn1czJ9eKKnOJ0z19CdbiZDNtBZNUBMJYTahqVrYxXEyCqmBDqg2p3hipPn6e6ztHpB8R0hOzfJda/9oGOYERgtscoKCwEdloQ6BVpXBrcaJBokEB0SPHEkmiaZvQrlSfHlupPDjchwgMJ1EJMDwErMbALcabFmJIYdhQa0OtN0StT9316oCQTW3AWUT7XH+i194y5NtgFH1x3CASQFa8Kg0BFx5jAwYM3yV6FTCSuItdpthTgjek2pDqbZFqZ5W2NSjELCGoybEnbw7MCDgiZP+qgwrKcmoT22YdmScp/JZ7dtVpgCkjUuCa6aZ0SFFLJytAwpOqyxhGXE6/2GJKvrwS5bivfx5PtMSqAvVRuXAtAOYNZTweQAea1CuzGyTXGjgeoQcPNdz3njMtvl22DD3H/fpuFuA2bttTzZo/6Fp+1GD8EfMDfj3IyMjkyRSdQHpjMaRXoo0oJZgkrRWU7sw2LBV4L8fwMmawqqS5w4bjGszewX4pw74lGJOs/ajuRLpwjsNPCCjLxnUNQyzF4hXFbeUoC/FC9SanM5p8Jrs13vdfIUOqQklXjdMLaVp8ujpXszGPJIzfdzM2F7q0dh3wWCqCbS1LNc4UAv3J0WGCcKwAXJuedDrpeBoapC2hIZpKTphcNwW0p3UzlKy/gnSq+iGWbc11hedycqOmuUt9sR/r/SDDsjjIeviFTuPoGw4fZLIJdpC9DKfk/hI+OcTo8y9Bfp669XHPClcD/6ALjV9E/lI5jy2GdLiGcmXj1ZMKPB/2xk7Fh1nAcV/BGewt1gcyhkOvlmzsi/PZCW5FNMCtWWqS5uevv3lRyds0Kthr3MFpMONQcsl9eO68dB69KcWbUrwpxZtS/LZK8aMXL7H0chtDpbgvvknq1/Bp2+f8ShmOQv5PHcZXV6apruKN0DrtbAdMR/JWVU0l4h1dvF9iMalfa5zpi6+RwLvpAdJQ6xden8V9f7tPbxeJFXdL8XSc2575/OAaIE8hlBZ/ZrXlE3RMZ4m/ticftlXbGm3H7ODbeETneQFi5mZlCPXu9rY0060SJUCxpbWs5Bb2afl1H/HAK7EH7L5UmPLEr23turikd9TunM4bo2/zXi81NnP+po5k/iCtixezJ7M8pdUrLj0rVslu2VyodtGtTslITWUrgkxegscOd/Yjy0BJ0T+kOIgzioM465ZKlEpBjLiZfFixENKBhV9o3WX3MKErme9nsUhNsKj7gREP/pA+wEwDU9RWUe9KhTq2rbl1sRune0xByHYjSdfwLFQ6GBhy3jo+algWPqTbFex8ER3j1qojyhkvd4yoSnIra7ANfpWOeiFSnQ76FUR3gLNda9wMkg8JFJEVFzFbM17w1TUzAzFB6Kchd3ek77vT9/f0OEGWun8+gRoSNi6oIpYgCywkxIhrIFSngrx1Hgv5hC7oaQ/48GiAUBxn7OU51OHJsZcdip8cn53j4GE6fiB64VMnb8gR/L+b4RfLweWSx8/uMi3NuKG1dTeLMunvP9nH6cc= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Rerank a list of documents based on their relevance to a query. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/resume-an-agent-turn-with-executed-tool-call-responses.api.mdx b/versioned_docs/version-v0.2.23/api/resume-an-agent-turn-with-executed-tool-call-responses.api.mdx new file mode 100644 index 0000000..4074c80 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/resume-an-agent-turn-with-executed-tool-call-responses.api.mdx @@ -0,0 +1,69 @@ +--- +id: resume-an-agent-turn-with-executed-tool-call-responses +title: "Resume an agent turn with executed tool call responses." +description: "Resume an agent turn with executed tool call responses." +sidebar_label: "Resume an agent turn with executed tool call responses." +hide_title: true +hide_table_of_contents: true +api: eJztfWtvG0mW5V8J8Mt0A7Jc05jZ2fF+UrtcvQbsKcNSe7DoKrCCzJCY7WQmJx9SEQX9941745GRDypJiVVOJk/DqBbzERnPEzfOuXHjt1muik2WFqqYvflt9pfvvqP/i1SxzONNGWfp7M3sStxUeSqyxT/VshTxrSjKXMm1iAvxg0wKdSGycqXyh7hQQqbiqtimy/elymWZ5SK7FVd3Ki0pic/2U9f8/ttVlX61qRaXs4vZMktL/SRlQG42SbyUlIHX/ywoF7/NiuVKrSX9VW43SufLvKpf3OTZRuVlbMpQ6i/N4yh4UOc3Tu/0g81y/T2N/6dSIo70R+PbWOXiVmdYF0VQEuIhLldxKqTQdVPQG48XM/vnC5PXBb1XecHFC1OP001Vztf6gryzZTFfkHkut/oDcanWfD1L1Y+3szf/GKqLPEtUX051Dgpd0bOqUDln/FZWSeNCWJKPVVGKhRI/8e2fZqLMXLG2ukC6H0j9T9BNYXNPxQnas5nejakDukn9g6rEvnYhHlbxciWWuh/F6TKpIt0Y6tdS96vI9DL9ZBRLncVOFdjCPV7sXz3m7u7qiddUlLB+6ithgb6P6dc6TrnLU2quXK6Y1HKX4ip5kNtC1yMn89OMG51THOzVVZ70jcy/f/7gPsUJCf39SJaSb+jeSzd0t1tLrmm684afe/0bfe7x/yxkof7Xv138RnceL8V/ZaWu8JV+mt5fZlUSiZW8VyJR6V25EokupBmsQ7mNhwcIdQP6jLktNlmse5T+Q/cuyrYGpqzKl7or6VqSURTTazL5FHzoluFHQ9j/VHGu9Jj8B3/5Z529uKSeP9PpUyVT8XY0NDXPu3SZRXTlzcxUSCev5rJQ9KCKbF1zReuu73reUD6bSb7nNHjktFK3FaKbUg8DLsEhNcBldP0qqAr+3ltTYA3P604Zr2yxwj7Lw+llg4jGb2MM+QsvGkKUihlBnN5wbyMYcZD0nArlzwT1SQk+XZ1l8Elbm/r9KCwp5dtW5/a/5JrStZ9b6zmQu+Rv+0IE0AxoBjQbQLM2WgDUXg5q9GHfLvpGnig9yqIwnWAaebYxC3wDvgHfYK3BWgOaAc3OAs1grX0ba+3x56ee8eTerz3k3p9+3Jgi/VncEDV4Gys9kPUfVUGdlxLTN5OtWGzFh0Supbgu5fIrjbyNLArx+epvwiZ+GaawllviIHO1zu45IW6Zq0/vPdRUZZWrS1CDgGfAM4zNk8NlGJtAM6AZjM1JgRqoQeAb8G0c+AZgg7UGNAOaTQPNYK2Nkxo8JO/snlg7CoYDpFD5R+tK2M269RYUt3m25lZg10N21lwSSoQOlpd7zHxDbpJlliXNTuIu9LtJ0u2dbpJ0UzivW6ZSZZK81GGV0qR0zHdc6kS76kcGnTH98647+Iq1BQWjilkNsxps9NOazmCjA82AZrDRJwVqYFSBb8C3ceAbgA3WGtAMaDYNNIO1Nh1G1VKKvdzqTZYlbiv+MMeaq40eoMqOWDNcdbtSa1kuM07vs6XjWn9+bCf1IS74abev3cBOnOoCyVIPAmYsacc9b7Av1eYoO9/3jgJA4PT+e9f56LVLl5FnvE+v2fdlrks3l71d1SC1vqKRRb0q43W3BShduuGTFTbJS0PnrjeJOvIHfKJ1Dewa7Cqt1tTf4vRW5RrOeMDozjBXv6plxV/QKazIc3dOnZFmbbXO8u08VzoNdS+TsEte62/dmGHXymQADJxHIvn1PwopYRsrmMWDzAQzub+qC7XWuJvMPfs+CHZDyoAsCt29ZdpEvObVfo3AP7NDKOCoCJTbfymaasEhURXa74PJhzUFawprw1Mzo7A2BJoBzbA2nBSogckHvgHfxoFvADZYa0AzoNk00AzW2jiZfCJVs808V9LGtN1Fq6o0mme3c2bFL9yvtefqs6rk29lXlRZUsmb+P3P64mG1rVlAQV/e6J56p1IKzau/dylMqAbdnXP1RrwSv1zrZ8zLl0EOfnkjbnwyt3EaF6tGOvwRYi/z2pP4sj85W4TBFBdV6X7qG1JspG5kWbtNi1evRFVUFEfiwskQRDNfcromEK3cOo9o6x9NMWSpTeO0Ut04vBTu1z8bcKbtkjTqvlGQXA99fZd7P93VpYjuVMlkOnPjlJ0nxY2nh+tOP3GXfspza8/KoNO/Frnul/NCyXy50pcfsuQ2l+u5TDYriq27WWVlpuufsSBSc47ascmV/m84jv5axYmuTdKTQhPK5YqkIJnfVWvdN/pXLAH33C56/xjfKw3K7Z3Kw0u62IvmlYXOtJJp86EkOWzFdexv//w4qqrQ2eF/QTPO/9mLXAehci1O1r027CktqfKt5GbZoSr6oV9cindyuQq2QpCOIlwKNtD45Ys2qDTxO8jnW6Oc6SSHBVVKTGrMs8DpNJo/eUHoz8H+jV1bWg6aBa0UWouaobjX0cQac5iVzkif6xaJYnTbB3aJc3usrCDUnp1Q28lAc2dVcAtTJ6ZOTJ3HnTpvwg2DBa3dzXhTtaXaOBTkyEPumLsaRzl22wUkystPW1QkZmseZMGeU191yzbdOuCiAZoMNBlI/9Pix0D6A82AZiD9JwVqcNEAvgHfxoFvADZYa0AzoNk00AzW2ihdNPpG00B0NNpVpkrZGgyHU9BtDr2XW24T0F3e2hih3emyZ2dmcFTHlc+lcIURckHeDJ61rDehvZyxHtiV2mkGd8NUfd++050kt6eym81mNePjKqmBVNUh0ltFfeckrn5p1eTR62CQVyGv7i+vNr9Xzw3hdV2w+zhLpEm/by+pv10PHJOA8bQatFb8+/NE3aseO+taX87jciv4vitdnauLp2o7I/1GGve4PM8aAs0Xl8QH/rAuKnmEef+5wa4c4KJ15yDriT0wts6OYh+zGiHrXJ/udNA3B5hD2MluKjZqGd/Gy6BfkF3FUpyI1KK6u6PHyMmuVInSKeTbw9C13WGCigyHiLxV5fZLXd/HA/DwK9zTST/dgc/BUAA4A5z3B+fOF0KE7twkmNY9J8vn0UK3dLFvUxceTvltXlTReojlfpu8cuesMcBzdcbaVOEeYe9AjgYlAEoABOfpcQEgOIFmQDMQnJMCNcjRwDfg2zjwDcAGaw1oBjSbBprBWjsZOdqRdo7Fi2p9pk32HVnebDKRPWxhUJiPzGR+dkTmDhbd8J3C8507qfQ9JojGltZglvDk73DvBh1/dnQ8YgY70ELMYFhhsMKwppyK+YU1JdAMaIY15aRADQoA8A34Ng58A7DBWgOaAc2mgWaw1kapADwiZjBiBgfNjsCHnS0u3z7aHwIfzk4h8CFiBp9IzOC2LAjlFsotgghjLsVcOpq51MdXQBBhBBEGbwbe7Fx5MxBmUAGAZkCzaaAZVAD4bADfgG9TxTcAG6w1oBnQbBpoBmttlD4bfaMJQYQRRHi0QYQb2h3EVoitCCmMkMLHmBwQUvjYIYW7Yx14DbxGlGFwBuAMwBmAAQUDCjQDmp0NmoEBhV4NfAO+TRXfAGyw1oBmQLNpoBmstZPRqx1pd1ZRhrkym2//mEdKZ1zDldk7rHvqUhUFyxc6lcJte4pEVOVmf25ccHrUkbKq3FRlj0yE0LFYKGBqxdSKhcIY51QsFIBmQDMsFCYFaqB1gW/At3HgG4AN1hrQDGg2DTSDtTZKWvcRoWMROjZodoS762xl+PYx3hDubnYK4e7OPHSsV7JkWepS+yH2TGRBmDrY6LDRwTicqnEOxgFoBjQD4zApUIM+BHwDvo0D3wBssNaAZkCzaaAZrLVvpA8NziOjGmXdctNXc+WOi+IzRnRJKZqQrYzLnpMyehzTa9bShNCJ12pnOJduYh/ff3zX6EiN5A6ioj2xW2chqIQrn263LtLgq1QT3cBAzReCgFM/xIkqCAvWKoqlTUeDhG0/Tqbl9f/ykD36ok5kvREPK5X6eEJioe6Y+j9G1J6giLs+58S+CxHf1jF9nr1fhTZ+ZPZHnAY7O4qZ2cNSzDp7PhqVGY56I7a2uzztLEls7s0OFVbD5LJWDfW1K2qzeCmutwWBGReIQOK1utc3XulaVHJN1Vroll73Rg1rAgG/1zUP3qdRfB9HlR5x/IQzCOquV4sH9qNDE/tGbpNMRt1vvTNZdzHB7HOhfmGywDNxz86Sp0o2GLyJNyBxQ9mWnLve4n7rVO90WTkoI/UI9zD/CB7m3/JBxoSRc+4l1OxuEmt8KIij17zcGEoWeUzpF4pqIlebLC/NoT9/RHCq/hzxNiqTIbcL6qB4YbsPbeKUqbOzIuV2U51oELyBCKlhmQ9DpqBrN7fQuRYI5xYasYQ5Lgoq7ZW7pg73yQ7Idrbtdc5fHXTNdMOCsakLArQXcNDoGOOIDNJqDcrGHYzLvnEZqcQMyvbkscwV2Su1vUbicqqnS7dZktMLi3wsFvGlawUu0ksWCbScDspf/h6Lhu+53geWC1wSO4hpgjbd9IDh+vtTHO3KfoLbOHzd3W6J5sJ7sS11KY+3Ot7RJGlrYfy7N4p3AGsOgsbVZ44El4YdDj7JAz2u4GwGZzM4m7GzWWcif6dnWJWbxd8rZ4AErmY1vcfsbEz25EbmBS/q7UM0cvga2UVlVTxta/EKlVe1ocFzK+OELxfVcqlUpKKu0fG2ynOCNPqWmdLpa42jSk12njPpBXjVKEtPNe6cDH2t7Qu+z9ILxkFxnpHZMh6FBuZLt3EaZgFMGZgyMGVgysCU+aNNmVAltNKgSfy4PKOjgIb4xk+2PfamHH2HmSbrGDzeYh0bd8A69rGOnHCkSj2wD3NAxHEoZ3ccSpiZwND3V0lcoo1R/qi1YbMVofrgDgd3uGeva8/YHe5bk0ST8oODcy/QDGgG595JgRq2YgHfgG/jwDcAG6w1oBnQbBpoBmsNofoQqg+h+uByAJeDibgctDHuzEP1HTQLDhzF1dLEGnOYlc76j9wibw+3bXfXYVvDKysItWcn1HYy0PL4q29h6sTUianzuFPnTejmVpiAC+zJU1uq3ufqdxhy+3sE1VM5O0XUs2JBj4x27LYLSJRXw72Q2ZoHSY5u93qhELXcOuCiAZoMNBlI/9Pix0D6A82AZiD9JwVqcNEAvgHfxoFvADZYa0AzoNk00AzW2ihdNPpGk+cd3ddu82zt2TzeVTbJkIVyQd4MnrWsN6G9nLF2XaXFV3+u96m1PVpsE3DVW4cPYk+X9dllO0huT2U3m81qxsdVUgOpqkOkt4r6zklc/dKqyaPXwSCvQl7dX15tfi/Ycx5c1wW7j7NEmvT79pL62/XAMQkYT6tBa8W/P0/Uveqxs6715Twut4Lvu9LVubp4qrYz0m+kcY/L86wh0HxxSXzgD+uikkeY958b7MoBLlp3DrKe2ANj6+wo9jGrEbLO9elOB31zQJwuk4qMOOFDNtf9guwqluJEpBbV3R09Rk52pUqUTiHfHoau7Q4TVGQ4ROStKrdf6vo+HoCHX+GeTvrpDnwOhgLAGeC8Pzh3vhAidOcmwbTuOVk+jxa6pXuD4vQ1tY9nY97mRRWth1jut8lbg/ZXY9CaaGvaVOEeYe9AjgYlAEoABOfpcQEgOIFmQDMQnJMCNcjRwDfg2zjwDcAGaw1oBjSbBprBWjsZOdqRdo7Fi2p9pk32HVnebDKRPWxhUJiPzGR+dkTmDhbd8J3C8507qfQ9JojGltZglvDk73DvBh1/dnQ8YgY70ELMYFhhsMKwppyK+YU1JdAMaIY15aRADQoA8A34Ng58A7DBWgOaAc2mgWaw1kapADwiZjBiBgfNjsCHnS0u3z7aHwIfzk4h8CFiBp9IzOC2LAjlFsotgghjLsVcOpq51MdXQBBhBBEGbwbe7Fx5MxBmUAGAZkCzaaAZVAD4bADfgG9TxTcAG6w1oBnQbBpoBmttlD4bfaMJQYQRRHi0QYQb2h3EVoitCCmMkMLHmBwQUvjYIYW7Yx14DbxGlGFwBuAMwBmAAQUDCjQDmp0NmoEBhV4NfAO+TRXfAGyw1oBmQLNpoBmstZPRqx1pd1ZRhh87AobdpKsHqCplnHjC0252ijipw8qvKzMtXWHDgreqw34xLO4VZfZG59VJ2VRgl8VPcptkMuqU315nUSXktklu4bwUXBW+Hmqxe58jDIPSPMGNc4GYtHelcwS7+61TvdMfZs2bOoh7mH8ED/Nv+SBjwvd5nG4q7hAeGsJ3A4BoXu6j1bkgYqEILnO1yajLMRbsq548sQWNapViRfDWbF2xxcEqU6PTuCw91TPo72sq8D7dgjPI1TPRHhE83uoUjTuH9osjam8H9x5dS+bPFybfCGgSpM4V6QTuJ7dw7rvgH4qLTfp3o4H8hbAkdTRsur0jELa08VysPP94YAhs89qFeFjFy5VYytSK1spMwCRIZ/pB+kAU06ZPqFdYQWAFAT7ktJYO4EOAZkAz8CGTAjWoV8A34Ns48A3ABmsNaAY0mwaawVobpXrlyL1fe8i9YNvNDVGDt7zDQv9RFdR5KbGUAkKLxVZ8SORaiutSLr/SyNvIohCfr/7mlLDLMAUKFL2gkbnO7jkhbpmrT+891FRllatLUIOAZ8AzjM2Tw2UYm0AzoBmMzUmBGqhB4BvwbRz4BmCDtQY0A5pNA81grY2TGjwk761zm8IBUqh8+Iwm7x/Proc7T2EanPmG3CQ5Xlyjk7gL/W6SdHunm2QrKNvFH3PixpPOmAOR8sCoYlbDrAYb/dSmM9joQDOgGWz0SYEaGFXgG/BtHPgGYIO1BjQDmk0DzWCtTYdR9ac39HCr4YkNwxxrrjZ6gCo7Ys1w1e1KrdV3nkM3IPmHuOCn3b52AztxqgskKZwHM5a8tf/RhCE4ys53ROA+vwjcYWaCmdxf1YVaa9ytT/cYBrshZUAWhe7eMm0iXvNqv0bgn9khFHBUBMrtvxRNteCQqArt98Hkw5qCNYW14amZUVgbAs2AZlgbTgrUwOQD34Bv48A3ABusNaAZ0GwaaAZrbZxMPpGq2WaeK1nQh3bTqiqN5tntnFnxC/dr7bn6rCr5dvZVpRytupn/z5y+eFhtaxZQ0Jc3uqfeqVTlkkbtpTChGnR3ztUb8Ur8cq2fMS9fBjn45Y248cncxmlcrBrpmOjcGgjy2pP4sj85W4TBFBdV6X7qG1JspG5kWbtNi1evRFVUFEfiwskQfJ4tp2sC0cqt84i2/tEUQ5baNE4r1Y3DS+F+/bMBZ9ouSaPuGwXJ9dCnk2Sp99NdXYroTpVMpgcnTD8hbjw9XHf6ibv0+ZDuvpVBp38tct0v54WS+XKlLz9kyW0u13OZbFYUW3ezyspM1z9jQaTmHLVjkyv933Ac/bWKE12bN+Y09U6uSAqS+V211n2jf8UScM/tou99wm43DcrtXfNE3e4Zu70H8TbP6h1ecR372z8/jqoqdHb4X9CM83/2ItcRjpave0pLqnxrD7ruVxWDM+HFO7lcBVshSEcRLgVhqvPAgxNaG1Sa+B3k862P7z8sqFJiUmOeBU6n0fzJC0J/DvZv7NrSctAsOHD6Q0sTa8xhVjrrP+WBYnTbB3afngyhFkJtW6jtZKC5syq4hakTUyemzuNOnTfhhkE+N9werVNbqv70kd9hyB1zV+Mox267gER5+WmLisRszYMs2HPqqznbJHDrgIsGaDLQZCD9T4sfA+kPNAOagfSfFKjBRQP4BnwbB74B2GCtAc2AZtNAM1hro3TR6BtNA9HRaFeZKmVrMBxOQbc59F5uuU1Ad3lrY4R2p8uenZnBUR1XPpfCFUbIBXkzeNay3oT2csZ6YFdqpxncDVP1fftOd5LcnspuNpvVjI+rpAZSVYdIbxX1nZO4+qVVk0evg0Fehby6v7za/F49N4TXdcHu4yyRJv2+vaT+dj1wTALG02rQWvHvzxN1r3rsrGt9OY/LreD7rnR1ri6equ2M9Btp3OPyPGsINF9cEh/4w7qo5BHm/ecGu3KAi9adg6wn9sDYOjuKfcxqhKxzfbrTQd8cYA5hJ7up2KhlfBsvg35BdhVLcSJSi+rujh4jJ7tSJUqnkG8PQ9d2hwkqMhwi8laV2y91fR8PwMOvcE8n/XQHPgdDAeAMcN4fnDtfCBG6c5NgWvecLJ9HC93Sxb5NXXg45bd5UUXrIZb7bfLKnbPGAM/VGWtThXuEvQM5GpQAKAEQnKfHBYDgBJoBzUBwTgrUIEcD34Bv48A3ABusNaAZ0GwaaAZr7WTkaEfaORYvqvWZNtl3ZHmzyUT2sIVBYT4yk/nZEZk7WHTDdwrPd+6k0veYIBpbWoNZwpO/w70bdPzZ0fGIGexACzGDYYXBCsOacirmF9aUQDOgGdaUkwI1KADAN+DbOPANwAZrDWgGNJsGmsFaG6UC8IiYwYgZHDQ7Ah92trh8+2h/CHw4O4XAh4gZfCIxg9uyIJRbKLcIIoy5FHPpaOZSH18BQYQRRBi8GXizc+XNQJhBBQCaAc2mgWZQAeCzAXwDvk0V3wBssNaAZkCzaaAZrLVR+mz0jSYEEUYQ4dEGEW5odxBbIbYipDBCCh9jckBI4WOHFO6OdeA18BpRhsEZgDMAZwAGFAwo0AxodjZoBgYUejXwDfg2VXwDsMFaA5oBzaaBZrDWTkavdqTdWUUZ5spsvv1jHimdcQ1XZu+w7qlLVRQsX+hUCrftKRJRlZv9uXHB6VFHyqpyU5U9MhFCx2KhgKkVUysWCmOcU7FQAJoBzbBQmBSogdYFvgHfxoFvADZYa0AzoNk00AzW2ihp3UeEjkXo2KDZEe6us5Xh28d4Q7i72SmEuzvz0LFeyZJlqUvth9gzkQVh6mCjw0YH43CqxjkYB6AZ0AyMw6RADfoQ8A34Ng58A7DBWgOaAc2mgWaw1r6RPjQ4j4xqlHXLTV/NlTsuis8Y0SWlaEK2Mi57TsrocUyvWUsTQideq53hXLqJfXz/8V2jIzWSO4iK9sRunYWgEq58ut26SIOvUk10AwM1XwgCTv0QJ6ogLFirKJY2HQ0Stv04mZbX/8tD9uiLOpH1RjysVOrjCYmFumPq/xhRe4Ii7vqcE/suRHxbx/R59n4V2viR2R9xGuzsKGZmD0sx6+z5aFRmOOqN2Nru8rSzJLG5NztUWA2Ty1o11NeuqM3ipbjeFgRmhxVI3euXfbRBykbYByllypuLlkh/W71BfZLbJJNRJ9v2ujlzh3K+9AKF4K8VXBbfX+t4ipd7GLtBfp/QsHkLEVe1bYu5a2/3W6d6pz9c2FL7h/lH8DD/lg8yJpSbcztTDflpqOd2Y1bacb8xPCyacNH0oCA4zdUmy0tzkA+3yvCEuW9grydOR6LWoF7F0o/t4IwAdV9/QfINyTtIvTV4nlCS9qUVhnZOkUjfaCZ/oX+/FN3esVVKWsXfDvBDN0mZ1y40UsXLlVjyCKeIeMpM8+Q2kOkHLVzPsIcK6xSsU8C6nNoCBawL0AxoBtZlUqAGjQz4BnwbB74B2GCtAc2AZtNAM1hrY91DtbRhpzq4EOoPRA3e8vEN+o+qoM6bGqkq2YrFVnxI5FqK61Iuv9LI28iiEJ+v/ubCbF2GKdBWogWNzHV2zwlxy1x9eu+hpiqrXF2CGgQ8A55hbJ4cLsPYBJoBzWBsTgrUQA0C34Bv48A3ABusNaAZ0GwaaAZrbZzU4CF5b0X2CAdIofLhKB4++D67Hu6O0/FSN0k+jL7RSdyFfjdJur3TTbJ14vvF7pBCBzis+ngr/J06dAmfGjzojOmfd92hcZ47GFXMapjVYKOf3HQGGx1oBjSDjT4pUAOjCnwDvo0D3wBssNaAZkCzaaAZrLXpMKo+unMPt0rRmF0cg2GONVcbPUBV6oPM61+6Xam1LJcZp/fZ0nGtFKi6mZQLEu32tRvYiVNdIA4s3zga1ISLeHME+3XfKAAETu+/d53Pxw1xp7Ae+D69dnmcWCGUG33DJytskpdHChTS/YBPtK6BXYPdxZeIUxuHhgYMBRI3B7/SF3QKK/Lc5Yj7NGvzebNzf95s2CXpiNobM+z6Q0L4PHbPpw1m8SAzwUzur+pCcaDvuWffB8EOR8/CmoI1BWsKa0OsDYFmQLNzRTOsDcHkA9+Ab1PFNwAbrDWgGdBsGmgGa22cTD6OnsXRs0+JG08PVxw9u18aOHr2eNnB0bPhBpXxHD170CxYh+q3omYo7nU0scYcZqUz0ue6RaIY3e4Yih3i3B4rKwi1ZyfUdjLQ3FkV3MLUiakTU+dxp86bcMNgYQ4QovGmakvVn0TyOwy5Y+5qHOXYbReQKC8/bVGRmK15kAV7Tn01Z5sEbh1w0QBNBpoMpP9p8WMg/YFmQDOQ/pMCNbhoAN+Ab+PANwAbrDWgGdBsGmgGa22ULhp9o2kgOhrtKlOlbA2GwynoNofeyy23Cegub22M0O502bMzMziq48rnUrjCCLkgbwbPWtab0F7OWA/sSu00g7thqr5v3+lOkttT2c1ms5rxcZXUQKrqEOmtor5zEle/tGry6HUwyKuQV/eXV5vfq+eG8Lou2H2cJdKk37eX1N+uB45JwHhaDVor/v15ou5Vj511rS/ncbkVfN+Vrs7VxVO1nZF+I417XJ5nDYHmi0viA39YF5U8wrz/3GBXDnDRunOQ9cQeGFtnR7GPWY2Qda5PdzromwPMIexkNxUbtYxv42XQL8iuYilORGpR3d3RY+RkV6pE6RTy7WHo2u4wQUWGQ0TeqnL7pa7v4wF4+BXu6aSf7sDnYCgAnAHO+4Nz5wshQnduEkzrnpPl82ihW7rYt6kLD6f8Ni+qaD3Ecr9NXrlz1hjguTpjbapwj7B3IEeDEgAlAILz9LgAEJxAM6AZCM5JgRrkaOAb8G0c+AZgg7UGNAOaTQPNYK2djBztSDvH4kW1PtMm+44sbzaZyB62MCjMR2YyPzsicweLbvhO4fnOnVT6HhNEY0trMEt48ne4d4OOPzs6HjGDHWghZjCsMFhhWFNOxfzCmhJoBjTDmnJSoAYFAPgGfBsHvgHYYK0BzYBm00AzWGujVAAeETMYMYODZkfgw84Wl28f7Q+BD2enEPgQMYNPJGZwWxaEcgvlFkGEMZdiLh3NXOrjKyCIMIIIgzcDb3auvBkIM6gAQDOg2TTQDCoAfDaAb8C3qeIbgA3WGtAMaDYNNIO1Nkqfjb7RhCDCCCI82iDCDe0OYivEVoQURkjhY0wOCCl87JDC3bEOvAZeI8owOANwBuAMwICCAQWaAc3OBs3AgEKvBr4B36aKbwA2WGtAM6DZNNAM1trJ6NWOtDurKMNcmc23f8wjpTOu4crsHdY9damKguULnUrhtj1FIqpysz83Ljg96khZVW6qskcmQuhYLBQwtWJqxUJhjHMqFgpAM6AZFgqTAjXQusA34Ns48A3ABmsNaAY0mwaawVobJa37iNCxCB0bNDvC3XW2Mnz7GG8Idzc7hXB3Zx461itZsix1qf0QeyayIEwdbHTY6GAcTtU4B+MANAOagXGYFKhBHwK+Ad/GgW8ANlhrQDOg2TTQDNbaN9KHBueRUY2ybrnpq7lyx0XxGSO6pBRNyFbGZc9JGT2O6TVraULoxGu1M5xLN7GP7z++a3SkRnIHUdGe2K2zEFTClU+3Wxdp8FWqiW5goOYLQcCpH+JEFYQFaxXF0qajQcK2HyfT8vp/ecgefVEnst6Ih5VKfTwhsVB3TP0fI2pPUMRdn3Ni34WIb+uYPs/er0IbPzL7I06DnR3FzOxhKWadPR+NygxHvRFb212edpYkNvdmhwqrYXJZq4b62hW1WbwU19uCwOywAql7/bKPNkjZCPsgpUx5c9ES6e+rBxnTkH9PZf4kt0kmo07e7XVz8A5lX9q3BFeV4M8WXCjfcevAipd72WmNrAfWGu8f4noenmaCNJ6QxIMU7fYk133cb53qnc5+YSvRP8w/gof5t6uLOdcFVbib1RofCgLrNS83xpaFIi6IHlFUxbnaZHlpTgH6I6JV9eeI91WZDLltUQcFENt9ihOnTL2fJSq3vepEo+INhEwNy/yCkR3uqXMt8NRAp81z19Th9hngdRS24XHdHi8nOUKDtFqDtHEH47RvnEYqMYO0tX5Jl7kig6Y26Eh9TvV86nZTcnphkY9FM750McFFeskqgtbbQfnL32NV8T3X+8B6gktiB3Wu5Np002D4fnsOpF3ZT5Afhy/M2y3RXJkvtqUu5fGWzzuaJG2tnH/3RvEeYs1B0Lj6zJHg0rDDwSd5oEsWvNHgjQZvNPZG60zk7/QMq3KzOnzlDJDAF63m/5i+jcm+3Mi84FW/fYhGDl8ju6isiqdtLV7C8rI3NHhuZZzw5aJaLpWKVNQ1Ot5WeU6QRt8yUzp9rXGWqcnOcya9AK8aZempxp2Toa+1fcH3WYLCODjQMzJbxiPhwHzpNk7DLIApA1MGpgxMGZgyf7QpE8qIVjs0iR+Xd3QU0BD/+Mm2x94UpO8w+7KQnvY7SRYyeLzFQjbugIXsYyE54UiVeqAf5rGI81PO7vyUMDOB4e+vkvhEO6n82WzDeILYfvCfg//cs9e5Z+w/961Jo0k5zsEbGGgGNIM38KRADXu3gG/At3HgG4AN1hrQDGg2DTSDtYbYfojth9h+cEGAC8JEXBDaGHfmsf0OmgUHzu5qaWKNOcxKZ/1ndJH3h9vnu+t0ruGVFYTasxNqOxloeQDWtzB1YurE1HncqfMmdHsrTIQG9uSpLVXvg/U7DLn9PYLqqZydIupZsaBHRjt22wUkyqvhbshszYMkx7d7vVCIWm4dcNEATQaaDKT/afFjIP2BZkAzkP6TAjW4aADfgG/jwDcAG6w1oBnQbBpoBmttlC4afaPJ847ua7d5tvZsHu8qm2RIQ7kgbwbPWtab0F7OWLuu0uKrP9f71NoeLbYJuOqtwwexp8v6sLMdJLenspvNZjXj4yqpgVTVIdJbRX3nJK5+adXk0etgkFchr+4vrza/F+w5D67rgt3HWSJN+n17Sf3teuCYBIyn1aC14t+fJ+pe9dhZ1/pyHpdbwfdd6epcXTxV2xnpN9K4x+V51hBovrgkPvCHdVHJI8z7zw125QAXrTsHWU/sgbF1dhT7mNUIWef6dKeDvjkgTpdJRUacKDZqGd/Gy6BfkF3FUpyI1KK6u6PHyMmuVInSKeTbw9C13WGCigyHiLxV5fZLXd/HA/DwK9zTST/dgc/BUAA4A5z3B+fOF0KE7twkmNY9J8vn0UK3dG+QnL6m9vFtzNu8qKL1EMv9Nnlr0P5qDFoTfU2bKtwj7B3I0aAEQAmA4Dw9LgAEJ9AMaAaCc1KgBjka+AZ8Gwe+AdhgrQHNgGbTQDNYaycjRzvSzrF4Ua3PtMm+I8ubTSayhy0MCvORmczPjsjcwaIbvlN4vnMnlb7HBNHY0hrMEp78He7doOPPjo5HzGAHWogZDCsMVhjWlFMxv7CmBJoBzbCmnBSoQQEAvgHfxoFvADZYa0AzoNk00AzW2igVgEfEDEbM4KDZEfiws8Xl20f7Q+DD2SkEPkTM4BOJGdyWBaHcQrlFEGHMpZhLRzOX+vgKCCKMIMLgzcCbnStvBsIMKgDQDGg2DTSDCgCfDeAb8G2q+AZgg7UGNAOaTQPNYK2N0mejbzQhiDCCCI82iHBDu4PYCrEVIYURUvgYkwNCCh87pHB3rAOvgdeIMgzOAJwBOAMwoGBAgWZAs7NBMzCg0KuBb8C3qeIbgA3WGtAMaDYNNIO1djJ6tSPtzirK8GNHwLCbdPUAVaWME0942s1OESd1WPl1ZaalK2xY8FZ12C+Gxb2izN7ovDopmwrssvhJbpNMRp3y2+ssqoTcNsktnJeCq8LXQy12m61c1FzMtw8P3KBoTxDlXDqToi2qY9vdb53qnc4FC+D15+2P4GH+LR9kTGA/j9NNxb3D40T4boAWzct9HDsXRCwUYWeuNhn1P18ZL9yPRlVMgSN4n7au5eJgyanRg1yWnuom9Pc1FXifPsIZ5OrZu3v4JjnJHhI83uokjTuH9pMjCnMH9yZdS+bPFybfiHYSpM4V6dTvJ/d37ssGDAXNJnG80UD+QliSOlQ23d4RJVvaYC9Wu388MD62ee1CPKzi5UosZWoVbWVmZ1KrM/0gfSCKaUcopC0sL7C8AFlyWusKkCVAM6AZyJJJgRqkLeAb8G0c+AZgg7UGNAOaTQPNYK2NUtpy5N6vPeResCfnhqjBW95+of+oCuq8lFhK0aLFYis+JHItxXUpl19p5G1kUYjPV39zMtllmAJFkV7QyFxn95wQt8zVp/ceaqqyytUlqEHAM+AZxubJ4TKMTaAZ0AzG5qRADdQg8A34Ng58A7DBWgOaAc2mgWaw1sZJDR6S99ahTuEAKVQ+fICTd55n18OdRzQNznxDbpIcTK7RSdyFfjdJur3TTbIVse3ijzmO40lnzIEwemBUMathVoONfmrTGWx0oBnQDDb6pEANjCrwDfg2DnwDsMFaA5oBzaaBZrDWpsOo+qMderjV8DiHYY41Vxs9QJUdsWa46nal1uo77KEbrfxDXPDTbl+7gZ041QWSFOuDGUve2v9owhAcZec7wnOfX3juMDPBTO6v6kKtNe7WR38Mg92QMiCLQndvmTYRr3m1XyPwz+wQCjgqAuX2X4qmWnBIVIX2+2DyYU3BmsLa8NTMKKwNgWZAM6wNJwVqYPKBb8C3ceAbgA3WGtAMaDYNNIO1Nk4mn0jVbDPPlSzoQ7tpVZVG8+x2zqz4hfu19lx9VpV8O/uqUg5l3cz/Z05fPKy2NQso6Msb3VPvVKpySaP2UphQDbo75+qNeCV+udbPmJcvgxz88kbc+GRu4zQuVo10TOhuDQR57Ul82Z+cLcJgiouqdD/1DSk2UjeyrN2mxatXoioqiiNx4WQIPuyW0zWBaOXWeURb/2iKIUttGqeV6sbhpXC//tmAM22XpFH3jYLkeujTMbPU++muLkV0p0oT0Lk+fvoJcePp4brTT9ylzyd4960MOv1rket+OS+UzJcrffkhS25zuZ7LZLOi2LqbVVZmuv4ZCyI156gdm1zp/4bj6K9VnOjavDFHrXdyRVKQzO+qte4b/SuWgHtuF33v43e7aVBu75rH7XYP4O09pbd5kO/wiuvY3/75cVRVobPD/4JmnP+zF7mOcO583VNaUuVbewp2v6oYHBgv3snlKtgKQTqKcCkIU50HnqrQ2qDSxO8gn2998P9hQZUSkxrzLHA6jeZPXhD6c7B/Y9eWloNmwYGjIVqaWGMOs9JZ/xEQFKPbPrD7aGUItRBq20JtJwPNnVXBLUydmDoxdR536rwJNwzyoeL23J3aUvWnkfwOQ+6YuxpHOXbbBSTKy09bVCRmax5kwZ5TX83ZJoFbB1w0QJOBJgPpf1r8GEh/oBnQDKT/pEANLhrAN+DbOPANwAZrDWgGNJsGmsFaG6WLRt9oGoiORrvKVClbg+FwCrrNofdyy20CustbGyO0O1327MwMjuq48rkUrjBCLsibwbOW9Sa0lzPWA7tSO83gbpiq79t3upPk9lR2s9msZnxcJTWQqjpEequo75zE1S+tmjx6HQzyKuTV/eXV5vfquSG8rgt2H2eJNOn37SX1t+uBYxIwnlaD1op/f56oe9VjZ13ry3lcbgXfd6Wrc3XxVG1npN9I4x6X51lDoPnikvjAH9ZFJY8w7z832JUDXLTuHGQ9sQfG1tlR7GNWI2Sd69OdDvrmAHMIO9lNxUYt49t4GfQLsqtYihORWlR3d/QYOdmVKlE6hXx7GLq2O0xQkeEQkbeq3H6p6/t4AB5+hXs66ac78DkYCgBngPP+4Nz5QojQnZsE07rnZPk8WuiWLvZt6sLDKb/NiypaD7Hcb5NX7pw1BniuzlibKtwj7B3I0aAEQAmA4Dw9LgAEJ9AMaAaCc1KgBjka+AZ8Gwe+AdhgrQHNgGbTQDNYaycjRzvSzrF4Ua3PtMm+I8ubTSayhy0MCvORmczPjsjcwaIbvlN4vnMnlb7HBNHY0hrMEp78He7doOPPjo5HzGAHWogZDCsMVhjWlFMxv7CmBJoBzbCmnBSoQQEAvgHfxoFvADZYa0AzoNk00AzW2igVgEfEDEbM4KDZEfiws8Xl20f7Q+DD2SkEPkTM4BOJGdyWBaHcQrlFEGHMpZhLRzOX+vgKCCKMIMLgzcCbnStvBsIMKgDQDGg2DTSDCgCfDeAb8G2q+AZgg7UGNAOaTQPNYK2N0mejbzQhiDCCCI82iHBDu4PYCrEVIYURUvgYkwNCCh87pHB3rAOvgdeIMgzOAJwBOAMwoGBAgWZAs7NBMzCg0KuBb8C3qeIbgA3WGtAMaDYNNIO1djJ6tSPtzirKMFdm8+0f80jpjGu4MnuHdU9dqqJg+UKnUrhtT5GIqtzsz40LTo86UlaVm6rskYkQOhYLBUytmFqxUBjjnIqFAtAMaIaFwqRADbQu8A34Ng58A7DBWgOaAc2mgWaw1kZJ6z4idCxCxwbNjnB3na0M3z7GG8LdzU4h3N2Zh471SpYsS11qP8SeiSwIUwcbHTY6GIdTNc7BOADNgGZgHCYFatCHgG/At3HgG4AN1hrQDGg2DTSDtfaN9KHBeWRUo6xbbvpqrtxxUXzGiC4pRROylXHZc1JGj2N6zVqaEDrxWu0M59JN7OP7j+8aHamR3EFUtCd26ywElXDl0+3WRRp8lWqiGxio+UIQcOqHOFEFYcFaRbG06WiQsO3HybS8/l8eskdf1ImsN+JhpVIfT0gs1B1T/8eI2hMUcdfnnNh3IeLbOqbPs/er0MaPzP6I02BnRzEze1iKWWfPR6Myw1FvxNZ2l6edJYnNvdmhwmqYXNaqob52RW0WL8X1tiAwO6xA6l6/7KMNUjbCPkgpU95ctET62+oN6pPcJpmMOtm2182ZO5TzpRcoBH+t4LL4/lrHUzRqJdWwfJAxwcqcK3Z4tghK8YSyzRuLuAFsC81dL3C/dap3OjuFrQv/MP8IHu7L5M/B5NRzuzFX7bjfGDQWY7hoeqgQyOZqk+WlOd6H22p4Gt033NcTZyZRG1FfY0HIdnvGhXoEvCD5hhAepN4aUk/oS/uSDUP7qUi6bzSTv9C/i4pu79hAJa0fgB32h26dMq9daPyKlyux5HFPcfKUmfzJmSDTD1oQn2FnFVYvWL2Aizm1ZQu4GKAZ0AxczKRADcoZ8A34Ng58A7DBWgOaAc2mgWaw1sa6s2ppg1F1cCFUJYgavOVDHfQfVUGdNzUCVrIVi634kMi1FNelXH6lkbeRRSE+X/3NBd+6DFOgDUYLGpnr7J4T4pa5+vTeQ01VVrm6BDUIeAY8w9g8OVyGsQk0A5rB2JwUqIEaBL4B38aBbwA2WGtAM6DZNNAM1to4qcFD8t6K9xEOkELlw7E9fEh+dj3cHb3jpW6SfER9o5O4C/1uknR7p5tk6xz4i92Bhg5wWPVRWPg7dUATPkt40BnTP++6Q+OUdzCqmNUwq8FGP7npDDY60AxoBht9UqAGRhX4BnwbB74B2GCtAc2AZtNAM1hr02FUfcznHm6VYjS76AbDHGuuNnqAqtSHnte/dLtSa1kuM07vs6XjWil8dTMpFzra7Ws3sBOnukAcbr5xYKgJIvHmCPbrvlEACJzef+86n48m4s5mPfB9eu3yOBFEKDf6hk9W2CQvjxQ+pPsBn2hdA7sGu4svEac2Og0NGAovbo6DpS/oFFbkuctx+GnW5lNo5/4U2rBL0sG1N2bY9YeE8HnsnlobzOJBZoKZ3F/VheLw33PPvg+CHQ6khTUFawrWFNaGWBsCzYBm54pmWBuCyQe+Ad+mim8ANlhrQDOg2TTQDNbaOJl8HEiLA2mfEjeeHq44kHa/NHAg7fGygwNpww0q4zmQ9qBZsA7gb0XNUNzraGKNOcxKZ6TPdYtEMbrd4RQ7xLk9VlYQas9OqO1koLmzKriFqRNTJ6bO406dN+GGwcIcK0TjTdWWqj+f5HcYcsfc1TjKsdsuIFFeftqiIjFb8yAL9pz6as42Cdw64KIBmgw0GUj/0+LHQPoDzYBmIP0nBWpw0QC+Ad/GgW8ANlhrQDOg2TTQDNbaKF00+kbTQHQ02lWmStkaDIdT0G0OvZdbbhPQXd7aGKHd6bJnZ2ZwVMeVz6VwhRFyQd4MnrWsN6G9nLEe2JXaaQZ3w1R9377TnSS3p7KbzWY14+MqqYFU1SHSW0V95ySufmnV5NHrYJBXIa/uL682v1fPDeF1XbD7OEukSb9vL6m/XQ8ck4DxtBq0Vvz780Tdqx4761pfzuNyK/i+K12dq4unajsj/UYa97g8zxoCzReXxAf+sC4qeYR5/7nBrhzgonXnIOuJPTC2zo5iH7MaIetcn+500DcHmEPYyW4qNmoZ38bLoF+QXcVSnIjUorq7o8fIya5UidIp5NvD0LXdYYKKDIeIvFXl9ktd38cD8PAr3NNJP92Bz8FQADgDnPcH584XQoTu3CSY1j0ny+fRQrd0sW9TFx5O+W1eVNF6iOV+m7xy56wxwHN1xtpU4R5h70COBiUASgAE5+lxASA4gWZAMxCckwI1yNHAN+DbOPANwAZrDWgGNJsGmsFaOxk52pF2jsWLan2mTfYdWd5sMpE9bGFQmI/MZH52ROYOFt3wncLznTup9D0miMaW1mCW8OTvcO8GHX92dDxiBjvQQsxgWGGwwrCmnIr5hTUl0AxohjXlpEANCgDwDfg2DnwDsMFaA5oBzaaBZrDWRqkAPCJmMGIGB82OwIedLS7fPtofAh/OTiHwIWIGn0jM4LYsCOUWyi2CCGMuxVw6mrnUx1dAEGEEEQZvBt7sXHkzEGZQAYBmQLNpoBlUAPhsAN+Ab1PFNwAbrDWgGdBsGmgGa22UPht9owlBhBFEeLRBhBvaHcRWiK0IKYyQwseYHBBS+NghhbtjHXgNvEaUYXAG4AzAGYABBQMKNAOanQ2agQGFXg18A75NFd8AbLDWgGZAs2mgGay1k9GrHWl3VlGGuTKbb/+YR0pnXMOV2Tuse+pSFQXLFzqVwm17ikRU5WZ/blxwetSRsqrcVGWPTITQsVgoYGrF1IqFwhjnVCwUgGZAMywUJgVqoHWBb8C3ceAbgA3WGtAMaDYNNIO1Nkpa9xGhYxE6Nmh2hLvrbGX49jHeEO5udgrh7s48dKxXsmRZ6lL7IfZMZEGYOtjosNHBOJyqcQ7GAWgGNAPjMClQgz4EfAO+jQPfAGyw1oBmQLNpoBmstW+kDw3OI6MaZd1y01dz5Y6L4jNGdEkpmpCtjMuekzJ6HNNr1tKE0InXamc4l25iH99/fNfoSI3kDqKiPbFbZyGohCufbrcu0uCrVBPdwEDNF4KAUz/EiSoIC9YqiqVNR4OEbT9OpuX1//KQPfqiTmS9EQ8rlfp4QmKh7pj6P0bUnqCIuz7nxL4LEd/WMX2evV+FNn5k9kecBjs7ipnZw1LMOns+GpUZjnojtra7PO0sSWzuzQ4VVsPkslYN9bUrarN4Ka63BYHZYQVS9/plH22QshH2QUqZ8uaiJdLfVw8ypiH/nsr8SW6TTEadvNvr5uAdyr60bwmuKsGfLbhQvuPWgRUvm5DWycY7ett9+qDibuxLTxWSU39GLT6Z6HWZK7l+u6rSr53KMvdIbK7rYulVnVoNtnuS/u277/p34FCOdM93BwHJJI54pMuEBpCKZo0dPNq8TWIT7fK1U9YKjQXr3rhyzalC9+KyCsUdJ/a1i/Z/b24+CfM0h3Ob1c06hLbvKOye4IdJZy9WWV6KolqvZb514Muh+fQwj5crktyo0HksKaCqsXbMff4Og2Ip42TfL5un6dNJluqyiVW1lukr3VSRXOhRGbzTyI4NalVKiv41+K0wYipPcaYwS0nwSN4EUSN81jrLSZkzgEhfrkME+gB62XJZ5WaWbObrkB5tW9i1lq+8oI9zNe2ovLwRXpXycPXp/aX4nFV3q2SrmydJsodCfP7hrfiP//3df/CcpX6V1OfD/qW7eiD5azz5bLp4nZ89e74dOX/5z/6Rs0xiGnkrsr7NnJrpl9OtS9gglbiL9TgXcp1V1pyIG7FnMagwqE5hUP3lPxteDuIj9XQ7sopwaP2/rDLrdfXrUilaS7KNr40ws3S/FJ/0SkNnqNS9R95JPUgSfTN3M9W/75qpCpXf68anFWpFFg3NfamoUvWrrm9yu1K2GjC0MLROaGj9ezhf8Uqc1qfXprfX37fD66rb5V1lRJfix0r3BG0X8qy0UHriSTM+zjKqN6db8qND5D2RMAYVBtVpDarv+r62xxDiUUJdTd4VVBpek/W6LOsOpZokillc+3AOtWNfvUb9Kf1vohekoGUeD1ITjJe7+S9uvTvn9e4vIqqYs9qoNKpXwVwv1vYs4ig8R+DCBGPQjzPF1u4LRbXQ8y9/0VAMvYcRiIypspXakpu1oF67JbZ9I3O93tToVLBGa3x3Z1x6x2fo3xtZrno5sDqss60w6ptUh5R23X/KvNL9qYMmNb3tv9ziUvb7tn3p5V+vOZ19P8095Jnf/dk8rE2tv2bRtuUB+RI4xjHLOGYZGuJpqBvQEOERMRbxEB4RQDOgGTwiJgVq8F8FvgHfxoFvADZYa0AzoNk00AzW2ijjm/SNJhyzfMrHLDfllppurxl40mjskYzkQBc0omuCduL/vVL67ZwVFH7HQat3sDtkBOw8jtkoSoEPoHFa4n7eVApMH1xl+udsk7FjE+sPb2av7//1NcsrxevfnDLz+NqKHq9/qyWTx9dUGXoGMzLG42tTTTNyUSXd2Wg8PGHOVmW5efP6tUy3r1b6ayp6lSRyLV8VpVx+vVxmxsAo1LKig3P5xe8dQv3jZ7pHqsTnWr94V8t1bf3hH4GUUMNPQPy3WX3PpNdPB+OTAcA1tK06PqeXPm3r/QMVRlxTYcS1VULdWb9UE6YT3P9rD1qQcFw0JFQZirphymwQaLS/j3UiP6Wi9T8pCsXmhdPuCnNo7UrFuUai3FSRVQE1jt3KJXVvSlTm6icK0JFkOct83cQX5OVGx9fmNDWZXH2kKB4FDwTqQmvJspGt42eKm+0aClSqM9RLLarQzPV6k8iYndatCWqG6z9MrzIDVv/xJhBT7VClqw2h00bYelPrj3bk6o5Oo5NS/e03Mob+niePj3RZj7uchqX+k50cFtTt9UDTSwL6O+q3rcLG+9Nniz9/FoNSbm+x3WyVEmFyL5OKfuk/v6rtLNCQaR47Qq56Rd5n5Cuo+CPlrKsBPyNbrul5jl/p7qaNBGpPc9faNq9ujKnl3u6IxFQi88bVcqk25ZPP/hzMOJ9+vL7RDy+sEE3hgPTVXD5QX9T/5ZxmXCts+vC132aJTO8qXqbNTJr0v/8PEEP6ZA== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Resume an agent turn with executed tool call responses. +When a Turn has the status `awaiting_input` due to pending input from client side tool calls, this endpoint can be used to submit the outputs from the tool calls once they are ready. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieve-an-agent-session-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/retrieve-an-agent-session-by-its-id.api.mdx new file mode 100644 index 0000000..9d0fbaa --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieve-an-agent-session-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieve-an-agent-session-by-its-id +title: "Retrieve an agent session by its ID." +description: "Retrieve an agent session by its ID." +sidebar_label: "Retrieve an agent session by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztXW1vGzcS/iuEvlwL2HJatNer75MvTdsASRPYboFDEqjULiWx2bcjd+0Ihv/7zQy5u1ztypJfikjKHIKrvdwdDoczD4cz5PhmZJQt8swqOzq9GX377Bn+J1Y2MroodZ6NTkdn4kJZCz+PR0ejKM9KlZX4liyKREcS3zr5y+KrNyMbLVQq8adyWSj4OJ/+paISPixMXihTateRdRQnOg7etaXR2Rze7fb/e6b/VymhY+hXz7QyYpYbUS6UAGaulLHEgvAkR7dHDfVMpmoz/V+rVGbHRslYThMl8KOmh4BoWZnMBtSkMXIJxHSpUrt5xPj5I4eLJMS1Lhc6E3JovH+PNHVWVOUkhQdyru6UQJ6pN7PR6btNsjB5MjgvwIEF1RpVVhlifCarpPMgHMnrypZiqsR7an4/EmVeD2sJA9JWSPgnsFF47nE4gQZ36V06GWCjyGckEv/Zkbhe6GghIpkJnUVJFcNkqE+lkFkscngRO4i1BBZ7IvCDuz3aXjyudb14dIpDCeXTPgkH9JPG31KdyRKnGKjV46qHiTM3FmfJtVxakCOReT+iSSeKG7W6MskQYPx+/qruiggJ6D+WpaQG0F5sALVLJUkaW07pvZMb7O7231Np1T+/O7rBltux+C0vQeALeBu/j/IqicVCXimRqGxeLkQCgywtwtMmbvVmA0E1wG5csyhyDRoFP4B2IduAl3llIlAlkJKMY42fyeRt0NFMJlYdAbL+r9JGgU2+o54/AHu6RM0fAX0UMg5vzUTj9LzIojzGJ6cjJ5Aer+6xUPiiir2sSdCg+rXmbeKzS/Il0SDLWaHuBQJTCWZAI7iPBGiMtV4FoqD+nrsBvwR17I3xzA8r1Fkyp8cZEdpvx4aaB48yIaTiLIjobdY2hJEakh4iUOomkCcSvFucZdCllyZ8H4cjRb69OJe/0SJad5fCqk8qebMtRDCaMZoxmm1As1W0YFB7PKhhx828QINJFFhZHNIJlpEHO7OMb4xvjG/srbG3xmjGaPZFoBl7a5/HW7v9cNc7TXDv00Bw76s3hRvS1+ISQ4MzrcCQ4YfKovIiMWhMlmK6FK8SmUpxUcroI1peIa0V52e/CE98HFJI5RJjkEal+RURopk5e/uygZqqrIwac2iQ4ZnhmZ3NvcNldjYZzRjN2Nk8KFDj0CDjG+PbbuAbAxt7a4xmjGaHgWbsre1maPA+vNPxxPagYGggVpnX/ihhn3V/WlDMTJ7SLNDRQzqsGSFKhAcsx1usfJuOSZZ5nnSVpH4wfEwSm9cek8RGUR8GplCqTJLHHlhFmkjH9VNTx7ArvLLxMGbzfq0OjWD9QDmiyqsar2rso+/XcsY+OqMZoxn76AcFahxRZXxjfNsNfGNgY2+N0YzR7DDQjL21w4mo+pDiYGz1Ms+Tcx/x2xxjNaoAA1XeYp25wrzibPlYps6u8qiOtX64XSX1Slt6u77X7mBHZzAgWYIRUMQSb9zTBftSFU9y833rKgAITi9/qpUPPxvXjDzge/zMfy8NjG4iB1XVITU8AWRRx6VO+zOAdLGhISs8ybEL56ZFop64g4ZoK4F1xq6yKkV909lMGYAzMhhQhon6pKKKegAKCzy5O0FlxFVbpblZTowCGupKJqFKXkBfl87sVpgMgIF4xCA//JujibnJClbxgJlgJW+ewqBSwN1k0kTfN4LdpsyAtBbUW2ZdxOs+Hc4RNO+sSRRQVQTk9h+2my24T1WF1e85ks/eFHtTvDfcNzeK94aMZoxmvDc8KFDjSD7jG+PbbuAbAxt7a4xmjGaHgWbsre1mJB+DqnkxMUr6Kr7rwqoqiyf5bEJR8aP6t7SJ1edVSc35R5VZHFmX/3OiL64XyzYKKLDnAjR1rjJlJFrtWLhSDaDORp2KY/HnBbzjPh4HHPx5Ki4bMjOdabvo0KFOMHpp2pPE42FyfggbKU6rsv4VGqQoJEyybI9Ni+NjUdkK60gc1WkIDDOPia4rRCuX9Ylofz4aa8jinOqsUv06vFjut3k3iJmujqQj+85ADJg+tJL2YyuMIp6rkoLpFBtHdu5MbtxtrmvPidf06yrM62K8rX5NDejlxCppogU8vs6TmZHpRCbFAmvrFou8zEH+hAWxmlDVjsIo+P/Qjv5T6QSkifmk0IWqucJUkDTzKgXdGN6xBLHn1aEP2/hWNJDbuTLhIxj2tPtkCkwrmXVfSpL77bieuu8PtzslCmCH/gXTOPlrELnuhcptcrLV2lBTVlKVzyVNy5qsYmP6dixeyGgRXIXAPIqoKQgnzvGjLqh08Tvg87nLnAHJzQlVJCYB8zxw1jmar5qE0NfB/Y11V1rutQr6VGib1AyTe72cWGcN86kzzM/1h4Q1uv0L65JzW+ysOFH7xSVqewx0b1YFTbx08tLJS+fTLp2X4YVBi3t3Z2+q9VQ7f6vkiU3uKW817qTtrg4QQ17NsoVDomjNtbR0cuojzGz3WAcf0eAwGYfJOOi/X/ExDvozmjGacdD/oECNj2gwvjG+7Qa+MbCxt8Zoxmh2GGjG3tpOHtEYsqYN1dHwVpkq5Yox3D8EvRpDH4wtrwag+3Fr54T2l8uBm5nBn+o4a7gU9WCEnOJphiZq2V5Ce3zEesOt1N401A1O9EP3TtcGuZtQdnfafM74aTOpQaqqF0hfGeqLOsU1nFp1PDZ5ME6vcnp1+/Rqt792bQifw8CudJ5IR3/oLmnT3BqOI+BOWm30VprvJ4m6UgN+1gU8NrpcCmqvR9dydXSXtHPM30h3PM6YvJOg+aMm8Yo6hqHiibDm/NxGVQ5w0R/nQO+JTmAsaz+Kzpi1CNlyvb/LwdAa4P4IO/pNtlCRnuko0Av0qygVJ2I1reZzfA0P2ZUqUUDBLO+HrqsKEwgyNBE5U+Xyj1beTwfgYS+k6Zg/XYPPgSkwODM4bw/OvR5ChO41IkyD5uRmEk9hpu22U20bOKWvaVOF+yFK93vyqv47awTwJE4NrgpphG/hdDSHBDgkwAHO/YsFcICT0YzRjAOcBwVqnI5mfGN82w18Y2Bjb43RjNHsMNCMvbW9SUfXQbs6ihe3+ZnVYN8Tpze7kciBaGEwmNcUyTyvA5lrougu3imaeOfaUPoWC0TnSmuwSjTB383azeH4Ly4czzWDa9DimsHshbEXxnvKQ3G/eE/JaMZoxnvKgwI1zgAwvjG+7Qa+MbCxt8Zoxmh2GGjG3tpOZgBuuWYw1wwOpp0LH/auuHz+an9c+HC0D4UPuWbwntQMXk0LcuaWM7dcRJjXUl5Ld2YtbeorcBFhLiLMcTOOm32pcTMOmHEWgNGM0eww0IyzAHxmg/GN8e1Q8Y2Bjb01RjNGs8NAM/bWdvLMxpA1cRFhLiK8s0WEO7k7TrZyspVLCnNJ4adYHLik8FOXFO7bOuM14zVXGeaYAccMOGbAEVCOgDKaMZp9MWjGEVDOVzO+Mb4dKr4xsLG3xmjGaHYYaMbe2t7kq+ug3RdVZZiE2f36jYkVMA5w5e4Og6ZGylpKXwAVW197ikVcGXc/V1uih4qUV2VRlQNpIi4dyxsFXlp5aeWNwi6uqbxRYDRjNOONwkGBGod1Gd8Y33YD3xjY2FtjNGM0Oww0Y29tJ8O6t1w6lkvHBtPO5e56Vxk+f403Lnc32odyd1946dgmkyXLEkbdmNgDkYXL1LGPzj46Rxz21TnniAOjGaMZRxwOCtQ4P8T4xvi2G/jGwMbeGqMZo9lhoBl7a58pP7RxHdkpK+uPG3s1qv5zUfQ3RmCkWE3IC2M88JcyBg6mt1FLV0JHp2ptOZc+sdcvX7/oKFKH3L1C0U1gt2UhEMJZQ7cviyzoFSXRLwzU/SAoOPWzTpRFLEhVrKWnAyDh54/IrJz6f3zJHngIRNJCXC9U1tQTElM1p9D/U1TtCYa4rrs62Xck9Kyt6fPg+yp48SP3v+gsuNlhR+4Oix317nx0hBlavUu2rqo83ixJPPfuhgplw2TUZg3h2RnOmY7ExdIimK3PV1BGEmhZtybTSpxHUWXwTgut7hpWHzeuv2viPXn6YzWRUfLeM9CRe/2Lz+PQ4NYK+cKPbK2ca95QVFuKmqD3u2fPhq/EIN8ggPov88hEx2R6MkHhqXjUuVID/maiXfnJkzrVZcE408FCb13shhGXVZhtqbNvq4P99fLyrXBvU321Ubt0bIK/F1gHT9DLmPi2i9yUwlZpKs2yRkOqlQezraMF5sBw0EZLrHDq3A/XTv2QmpZSJ9v27N7GrpM8g7GJRZXK7BiUKJZTmL7gmw47vspUKbEc18a+whKmtOa4wUQS8QrT+3GnnlWaG0yVOWPAntuafU1FO29juGx1+bqX3rsZrmerEV6g4CSmNcIznXqnyMPZ25djcZ5X80WyhOlJkvzaivOfn4sf/vXsB0IR9UkiSIb6Baoe5OBlLM6dirf8bKn53nK+/XHYcqJE43q2QHfYLXI5fJwta8LWpSfn+gpwRaZ55dd33SkGy0bFRrUPRvXtj51jB+I1arq3LBua1n/zyi3b6lOkFG7uyOmGVdTtpcfiLbj+wFAJ2iPnEowkgUZTr1Tfr1uprDJXMPm4Zaxw3cOTT5moMvUJ5I3noJQXA5sWm9Yemdb34XpFW2PcMF44bW/79+Z11lf5xkEeizcVaIKSKa1KUwULT5bT35eM29viPhrRi6zdQZiNio1qv4zq2VBvW5gQWQmqmpxbHA3tpgbPEHu5NnGNemc2XQoNvt/LnzBmXEgD2z4waUuZRncCdXVfDk8KWS4GYzlBeWJPHyaVTs2GUi9NBbPQs8E2Stt0Tbzer2MftqFuO5yALj2UDR+jsDUb4EaY5V0q3pxnxDgDFvIFfmY6Acl2WJouuxx5/eyxtO4AYMMqhulg4hY5EBnBwGkuQVSno5Orb05IJPbkppbm7Yln4OSmndtb2vwjirvJp3zGaFGWxenJCXhPx4scdujxcZLIVB6D4kYfx1Hu8j9WgTrqckkf/lRD9rsP2EaV35Flr9+v8Htxgd+LC2/KdfV47NzJ8uqbgWlG5LMdDJAhKoWUKRwDWH6lgcj7TKz8D2BW0QSpLKaAq3Vl0BdKG8Bv48yaqqhT0GImI+VjPNKo93jkM8kNYVaf+BS3aVgQ3WCw3nH1Gs+FWjL/AuQIAIsy8eq1pXV2xBEc89z2e680GO0+KRLwZKnOvktbOW155+Tu9AV+OA3szzaxntMAEGCCUS/w05sbzJL8bpLbW3zsbAQVItYWF5J4OM0SDuShFjQ4so9q2TVd2DNX+BqpLC2gU9TI+3F47o31a7EJ7waZqg06W4YM1cwGgkX0eQK2NqHhA5hsdILOki/ATQB/CKXoms+iSBVl8GHP/+qA1S8vLgG//g/6iqMj +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieve an agent session by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieve-an-agent-step-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/retrieve-an-agent-step-by-its-id.api.mdx new file mode 100644 index 0000000..2ffd370 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieve-an-agent-step-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieve-an-agent-step-by-its-id +title: "Retrieve an agent step by its ID." +description: "Retrieve an agent step by its ID." +sidebar_label: "Retrieve an agent step by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztXFFv2zgS/iuEX64FEjtb7N7e5p5ybe+uQLstknSBQxt4aYm22ZVEL0klMQz/95shKYmylNhuDWySzqLYJBI5nBnOfBzOUFwNtDALVRhhBqerwYuTE/yRCpNoubBSFYPTwVnBzmaisBdWLM5D6+HgaJCowsJj7MAXi0wmHDuMvhjstRqYZC5yjr/Z5UIAHTX5IhILHRdaLYS20o9pgCz+VIV4Px2cftrW3pa6GMs0ImyslsUMGrb5vpwL9uYVU1Nm4TfsNhysj9x4X9Efu4X+XFuRjrntIzFVOsc3g5RbcWxlLnrp4ouaLAskHflE5YtMHHiAmmijAU+3S18UZQ6zMJDFVGhRJEjeKpWNxa1ISjcCUJhLkaXjhGcZ/JWLXOnlWAugIa55NriCLtJmSBeN5hKH6DAJD2PdMlkwDv/Q0sJkORMzKGvMTCqmvMzaT0GoXKUiG1fWvN3stMp65a+G5MZImJfCtoZsP43leVcayyaCfW7afB4wq5hMQSI5XYKk0jBunMSO278ZVvO7bvlTdz7Dy0pjnf5HXf8JIq2P9vCtO4yinocc5qc9B/WTmONXEv/KZcGt0sxGU13JIa3Ih+wsu+FLAzpzZD4PUA2e4tb5K3XWA1Xs4/nbaihHiMH44CncvQATwxfehbAZvjl17UYrHG79zwk34u8/Hq3wzXrIflUW3QhaY/9ElVnK5vxasEwUMztnGQhpDZrqNm7lboCDw/jXbKEkGA78AkaEbMNcq1KjtYOWeJpK7MazD9FAU54ZcQSY/mcptUjRj3HkyB+BPioZxbtjonF6XheJSvHJ6cArpMOrf8wENhRp0LVTNFh4ZXnb+GyTfONoQHfONqgHhcBUAkI4CfbRgPXw4+0qUoUb76UX+A2YY0fGsyBWbLPOnb7Niay4bYNK/eCbXAipeA9y9LZbG7SqaH2VQt0wkT6R4P3qtNGQQZvQP40lRb6DOpe/8hzphuFyCDKcSa52hQhCM0IzQrMtaLaJFgRq3w5qOHA9L/BCZwK8LI3pRMsI15ovcUrhudlrJ0T4RvhG+EbRGkVrhGaEZt8DmlG09tdEa+ur+9r4pKpajLXgIft9V1pVFOlYTceY4HRP3V+5MMZDmSqte63+EIVBydr8nzv67Ga+bLKADEdegKXORCE0R68dsveuB5izFqfsmP1+AW1852HEwe+n7LImM5WFNPMWHTcIZi+1qDONw35yQYStFCelrf6EF5wtOEwyz2ry7PiYlabkWbY8gteYeWaYZh46uqURmuV86R75tD424EXq5lQWpajs81po4woS7EYCWFZto5zppiQt3bcE0eD68NZZP74FKdKZsC6Z7nLjyE7keN2Q/n53xe69JYmKfuHW1p6dQce+JhrscmwE18kcHt+obKp5PubZYs5x3LmyCvTvsCAVY4m2vNAC/h/70b9KmYE2L2HsOISquLpCV9WzMgfb6N+xRLnnTdH7fXwnGsjtDDiNHoHYk/aTCTAteNFulGX77bgOPfbV+kGpAthx/6JpHH/pRa69ULmy49hqY0uJkRoavORuWjYw7q00wdOC65she82TefOAYR2FVRSYV+dwP1Zd+aepubTxO+Lzpa+cAcl3NUhvrikB+xwEccC8AJxVjeZZXRB63sAcVrtYgsFdjFV7ClGVQpuiZlzc69TEWmtYKJ1hfa4rUsHq2tpdxbkddlZUqP3uCrUdBqIQs/2Klk5aOmnpPOzSeVkFxM6pcO/u/U00kWrruM+BXW6TnY+F/BMC8nAAQkLoPsV9ZMykPxTRrIoGmzxY390UEFNe9bKFIrlszQ03gJHXsFFIN4510BENSpNRmoyS/o8rP0ZJf0IzQjNK+j8pUKMjGoRvhG8PA98I2ChaIzQjNHsaaEbR2oM8otHnTXXesRptqlVeZ/PcV2XC8g1n2D8FvZlD780tbyagu3lrH4R2l0uXjG4L98yf/+DZc3ZWc8kqYRif4GmGOmvZfIT27RnrylQ28tXnzXdqmydawhQ41YcDH5g9TaqC6J1J7jqV3Z62UDM+bCU1KlV1Eukbor6uSlz9pVXPY10Ho/IqlVd3L6+2x2vWhvg5CHYtVcY9/b5vSevXjeN4Av6k1dZope4/zsS16ImzLuCxlnbJ3PtKuoaro/u0rbB+w/3xOK1Vq0DzW0XirRsYRMUTYfX5ua2mHOFiOM6B0ZM7gbGs4ih3xqxByIbrx7sc9K0BskiyEoM4ZhYikVOZRHaBcZUrxbFUTMrZDJvhITsrMgEU9HI/dN00mEiRsYvwqbDL3xp9Hw7A41GcpWP99A58jlyBwJnAeXdw7owQI3TnJcI0WI7S43QCM212nWpTw6nr7TZVuB9y5f5APgS0tz6gdeqUEKo4iwhvqBxNKQFKCVCC8/HlAijBSWhGaEYJzicFalSOJnwjfHsY+EbARtEaoRmh2dNAM4rWHk05ukraVVm8tKnPbCb7DlzebGcie7KFkTDvXCbzvEpk3pFF9/lOVuc770yl77BAtD5pjVaJOvm73bopHf/dpePpzuAKtOjOYIrCKAqjPeVTCb9oT0loRmhGe8onBWpUASB8I3x7GPhGwEbRGqEZodnTQDOK1h5kBWBNdwbTncHRtNPFh51PXP762/7o4sPBY7j4kO4MfiR3Bm+WBalyS5VbukSY1lJaSx/MWlrfr0CXCNMlwpQ3o7zZ95o3o4QZVQEIzQjNngaaURWAzmwQvhG+PVV8I2CjaI3QjNDsaaAZRWsP8sxGnzfRJcJ0ifCDvUS4VbujYisVW+lKYbpS+BCLA10pfOgrhbu+TnhNeE23DFPOgHIGlDOgDChlQAnNCM2+GzSjDCjVqwnfCN+eKr4RsFG0RmhGaPY00IyitUdTr66Sdt/VLcPr3gpwlQv23bynFmlUV02F5TIz+6kBicWinCEjKMD2+nX0NXAYGi2bN/UUL1RIwDupfjw56S/JIUvC1F8S8UymiBY5zxDrRTpofVUEy2Imfbl8VH2JZpK5yHsLU22vNpbbMk4tVx/Hbcr538vLD8y3dvWgQWPh21zyNdbtmGuMdzyYudKghzLPOcx/8H5X22M3c5nM8SMwFFpLjicyPEr6924cZw6o4F1H9q1x6EwVIBublzkvjrXgKZ9kAPNNnxY7IStuefue6u3FRL9se2ESYH7iyohpK/8Oxo9fhfu1G0duaoy1xagkKbX/KrvN134m7Wa4mq1aeZGVOzXdoTzdOp+BPJx9eDNk56qczbMlTE+WqRvDzv/9kv38j5OfXSlB3HJ0zti+wNSjz+x4ys69iTf87Gj5wXNe/NLvOUkm0cvmuGo7DFEKOhfLirDxX+LP5LWAn7kqwx3LsnV4hZyKnOoxONWLX1pHj9g7tPTgWSZ2rf+p0sf54jYRAmNQF4FzWD9dyD9kHyD6AIYsWA+fwTrGMnipq5Xqp7tWKogCrmHyMbItMYbBS34KVhbiFvSNV/6IoAZyLXKtR+RaP8XrlYvO8bjMhbf2ZvzgXmddk6+UkQ7Z+xIsQfDcrUoTAQtPodz38GnlXvWmqZMAuIcwORU51eNyqpO+0XZwIeclaGp8ZlAatyvrvS4v6LXeQrrN4WQJu3LD3rzCvNaCa57jvQ/GVUP8DRMD19rveiVSWnA7792BNyepwh5VsZmwzfEfmEocpdG91SXMRccTm5RSzQNs2g0eituLi9DpkHw0qYBdmcAeB9VEnYLYWQ04oudgv0Gv3OHDuYKmA+jsDARGOx2Mrn8YuSk2o1VlHetRUPdo1UzWeoTij1ZBbdAGeIEGXoY15lHcquHtzaV5B3NrF6ejEURrx3MFLdPjLOM5PwZHSf4YJsqnxY0A85d26Tq+qpaIT1f4zp2MRXmCP73F/uwC+7OLAB3V6Voc3Cvt+ocePSLSmhbm8BgFY8oucwxrx7UEIp8LtvEfwLpw2ylRpC7ba/wx0bmQGtYL7WHEnTJ1F7BMeYKn0pAo1+Iz5k8ypR1GdolPcFuIB0Y15jA9V+/wwjPj4GYBegRAR50EK9oFDVq6WDXr6U6dgy1h+mu0yCBmdieQfR7fW9Enr3FvR/DLaYQzwYLwacvzw1Whp+2MnGsW3AKmH60Gya9WmOz7qLP1Gh9D4K/RXOBXt3RO0DbAeFJp8Pe0P1Mdy/3sPLjOc7YH3PVqoipFF1iIhp18iX/Br3+IZYy46PcH4G8LEH4Fh9GsHIjH+3DyKxisLORQGmxj6NdoLBioO3M+h8gKQkg0P//2LEnEwkb9OiFrC43/8/oS8Pn/K8d9KA== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieve an agent step by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieve-an-agent-turn-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/retrieve-an-agent-turn-by-its-id.api.mdx new file mode 100644 index 0000000..c3ea4e4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieve-an-agent-turn-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieve-an-agent-turn-by-its-id +title: "Retrieve an agent turn by its ID." +description: "Retrieve an agent turn by its ID." +sidebar_label: "Retrieve an agent turn by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztXW1vGzcS/iuEvlwL2FJatNer75MvSXsG4iaw3QKHJFCpXUpis28luXYEQ//9ZkjuLle7imRbQCR5CqOR9mXIGc48HM6Qo/uBErrIMy304Ox+8P2LF/hPLHSkZGFkng3OBufsplTZcHAyiPLMiMzgI7woEhlxfGT0l8bn7gc6mouU4yezKAS8mU/+EpGBFwuVF0IZ6VoxQG4s4+BBbZTMZvBgu+XfM/l3KZiMoVE5lUKxaa6YmQuGJNidNHOZMc6g9xrfWJ4M/McnkgdGb4XSlr2QusyK0oxTuMBnnhfXAleKL6ABaURqr+eZeDsdnL3fJAuVJ6Kvp9ADDYIelFoo2/EpL5PWhZCTy1IbNhHsg739YcBMXrG1AIakZhz+GN5kvvfITjCebXo3TgZ4k+VTKxL/2gm7m8toziKeMZlFSRnDYIjPhvEsZjk8iA3EkkMXOyLwzC1PthePu7tePDJFVkL5NFdChl5J/JbKjBscYqBW8VWxiSM3ZOfJHV9okKMl82FgB91S3KjVpUr6bOf3qzdVU5YQg/Zjbri9AdqLN0DtUm4ljXfO7HOje2xu+e8J1+KfP5zc453lkP2WGxD4HJ7G96O8TGI257eCJSKbmTlLgEmj0Vg39VZuNhBUA2zG3WZFLkGj4ANoF3YboCMvVQSqBFLicSzxNZ68Cxqa8kSLEwCZv0upBNjke9vyR+ieNKj5A6CPQkb21gw0Ds/rLMpjvHI2cALp9NVdZgIfFLGXtRU0qH6leZv62SZ5YWlYy1mh7gUCQwlmYDl4iAQsj5VeBaKw7b10DF+AOnZ4PPdshTprzelpRoT227Kh+sKTTAipOAuy9DZrG8JIBUmPEahtJpAnEvyyOE3QpJcmvB+HnGK/vTgXv/EU6frmUpgDrUrebwsRhGaEZoRmG9BsFS0I1J4OathwPS5wQyUCrCwO6QTTyKOdWcI3wjfCN/LWyFsjNCM0exZoRt7a1/HWlh+/9Ewd3PvcE9z75m3hWPqW3WBocCoFGDJ8KDUqLxKDm8mCTRbsTcJTzq4Njz6h5RVca3Z1/ivzxIchhZQvMAapRJrfWkJ2ZM7fXdRQU5pSiSGFBgmeCZ7J2Tw4XCZnk9CM0IyczaMCNQoNEr4Rvu0HvhGwkbdGaEZodhxoRt7afoYGH9J3uz2x2SgYGogW6tJvJex23e8WZFOVp3YU7NZDu1kzQpQIN1gOt5j5Nm2TNHmetJWkutC/TRJvr90miTdZtS/WhlJ5kjx1wyrSRDqunYo6hl3hkY2bMevnK3WoBesZpYgqzWo0q5GPfljTGfnohGaEZuSjHxWoUUSV8I3wbT/wjYCNvDVCM0Kz40Az8taOJ6LqQ4q9sdWbPE+ufMRvc4xViQIMVHiLdeYK44qj5WOZMrvNoyrW+nG5SuqN1Pbp6ly7gx2ZAUPcgBHYiCWeuLcH7I0odnLyfesqAAhOF68q5TO2FoHvyCPex9f8+1wBd2Peq6oOqeEKIIs4NTLtjgDSxRs1WeZJDl04Ny0SseMGaqKNBNYZu8jKFPVNZlOhAM6swYAyjMVnEZW2BaAwx527Y1RGnLVFmqvFWAmgIW55EqrkNbR148xupZMBMNg+YpAf/mZoYqYpHOFm8aAzwUxeXwWmUsDdZFxH3zeC3abMANca1JtnbcRrX+3PEdTPrEkU2KoI2Nt/6Ha24CFVFVbfp0g+eVPkTdHa8NDcKFobEpoRmtHa8KhAjSL5hG+Eb/uBbwRs5K0RmhGaHQeakbe2n5F8DKrmxVgJ7mvarguriiwe59OxjYqfVN/SOlafl8bezj+JTCNn7f5fWfrsbr5oooAMWy5AU2ciE4qj1Q6ZK9UA6qzEGTtlf17DM+7lYdCDP8/YTU1mKjOp5y06thGMXqpmJ/Gwn5xnYSPFSWmqr3CDs4LDIPNm2zQ7PWWlLrGOxEmVhsAw89DSdYVo+aLaEe33R2MNWRxTmZWiW4cXy/3WzwYx01VOWrJvMaLA9OGu1X68C1zEM2FsMN3GxrE7X0xufNlc1+4Tr+hndm7tWRl09GuiQC/HWnAVzeHyXZ5MFU/HPCnmWFu3mOcmB/lbLIjF2FbtKJSA/4d29J9SJiBNzCeFLlTVK0wFcTUrU9CN/hVLEHteZb3fxreigb2dCRVeArYn7SsT6LTgWfuhJHnYimvXbX9c7pUooDv2LxjG8V+9yPUgVG6Sk43Whpqykqp8ye2wrMkq1qavh+w1j+bBUQjMo7CKAnPiHD7pgEobv4N+vnSZMyC5OaGKxDhgngfOKkfzTZ0Q+jY4v7HuSMuDZkGfCm2SmmFyr5MTa81hPnWG+bkuS1ij2z+wLjm3xcqKErXPLlHb6UD7ZFVwi6ZOmjpp6tzt1HkTHhjUuHZ39iYaT7X1sx07NrldnmrcS9tdZRBDXvW0hSzZaM0d13bn1CcY2fa2DtqiQWEyCpNR0P+w4mMU9Cc0IzSjoP9RgRpt0SB8I3zbD3wjYCNvjdCM0Ow40Iy8tb3cotFnTRuqo+GpMmH4ijE8PAS9GkPvjS2vBqC7cWvnhHany56TmcFPdZzXvWQVM4xPcDdDHbVsDqE9PWK94VRqZxiqG070fedO1wa561B2e9h8zni3mdQgVdUJpK+w+rpKcfWnVl0f6zwYpVcpvbp9erXdXjM3hNeBsVuZJ9zR7ztLWt9uDMcRcDutNnor9fvjRNyKHj/rGi4raRbM3q+4a3p18iVp55i/4W57nFJ5K0HzR0XijW0YWMUdYfX+uY2qHOCi386B3pPdgbGo/Ci7x6xByKbXhzsd9M0B7kfY0W/ShYjkVEaBXqBfZVNxLBaTcjbDx3CTnRGJAApq8TB0XVWYQJChifCpMIs/GnnvDsDDVqymY/50DT4HpkDgTOC8PTh3WggRunMTYRo0J1fjeAIjrbcdal3DqX3bLqpwPWTT/Z68qH5nzQK8FacEV8VqhL9D6WgKCVBIgAKchxcLoAAnoRmhGQU4jwrUKB1N+Eb4th/4RsBG3hqhGaHZcaAZeWsHk46ugnZVFC9u8jOrwb4dpzfbkcieaGHAzKWNZF5Vgcw1UXQX72R1vHNtKH2LCaJ1pDWYJerg72btpnD8swvHU83gCrSoZjB5YeSF0ZryWNwvWlMSmhGa0ZryqECNMgCEb4Rv+4FvBGzkrRGaEZodB5qRt7aXGYAl1QymmsHBsFPhw84Rl69f7Y8KHw4OofAh1Qw+kJrBq2lBytxS5paKCNNcSnPp3syldX0FKiJMRYQpbkZxs+caN6OAGWUBCM0IzY4DzSgLQHs2CN8I344V3wjYyFsjNCM0Ow40I29tL/ds9FkTFRGmIsJ7W0S4lbujZCslW6mkMJUU3sXkQCWFd11SuGvrhNeE11RlmGIGFDOgmAFFQCkCSmhGaPZs0IwioJSvJnwjfDtWfCNgI2+N0IzQ7DjQjLy1g8lXV0G7Z1Vl2Aqz/fZbFQvoOMCVOzsMmhoJrW36Aqjo6thTzOJSufO5Ult6qEh5aYrS9KSJqHQsLRRoaqWplRYK+zin0kKB0IzQjBYKRwVqFNYlfCN82w98I2Ajb43QjNDsONCMvLW9DOsuqXQslY4Nhp3K3XWOMnz9Gm9U7m5wCOXunnnp2DqTxY0BrmsTeySyUJk68tHJR6eIw6E65xRxIDQjNKOIw1GBGuWHCN8I3/YD3wjYyFsjNCM0Ow40I2/tK+WHNs4je2VlXb6xVSWqn4uyvzECnGI1IS+MYc8vZfRsTG+ilq6EjkzF2nIuXWKXF5evW4rUIvegUHQd2G26EAjhvKbblUUWtIqS6BYGar8QFJz6RSZCIxakIpbc0wGQ8ONnyazs+n96yR64CETSgt3NRVbXE2ITMbOh/11U7QlYXNdclew7YXLa1PR59HkVPPiR+y8yC0526IE7w6IHnTMfLWGGVu+SrasqjydLEt97d0LFZsN41GQN4do5jpmM2PVCI5j5kys/vHjRf04D2QH5VD8XwxMZW33gCYpZxIPWOQ9wghLpaiKOqvyLBo1Je6uPtQEFeDVlmAKoUkKrnP735uYdc0/bol+DBs822eRrLM7G7MOYjdXzXBmmyzTlalGZqC3gBsogozkmZpBpJTmW3XRzortv27GmY7hMtm3ZPY1NJ3kGvLF5mfLsVAke8wmMXfBOqzu+9JHhWCNqY1thXU0LhI6ZiKMRYc45bhVZSnOF+RtnNthyU0iuLrOWR1GpHJa2+/UQc/AjXI1WLbxAta2Y1ghPtYpwYh/O310M2VVezubJAoYnSfI7za5+ecl++teLnyyyic8cLTfUL1D1IDHMY3blVLzpz5aa7y3n+5/7LSdKJILsHH00h7w5vJwtKsLa5cxm8hZgh6d56Scd2apQSkZFRnUIRvX9z61cOLtETfeWpUPT+l9eulWd+BwJgSsO6wnCVO0WeEP2DvxR6JAB7eEzDkaSwE1VzVQ/rpuptFC3MPi4jilx3sPtOBkrM/EZ5I2bc4QXA5kWmdYBmdaP4Xxl12u4irl22t60783rvKvylTDiIXtbgiYIntpZaSJg4sly+6OHcXOE2S+RO+GeLxAmoyKjOiyjetHX2hYmZK0EVY3PNHJjV1O9G1u9XFuLbTZZMAmO38UrjGIWXPEUdypqm/tyeyIH9ulqnQjfC27mvbGFplyubyBnM2GCBWyusJVG9kaVMBYdS2wCiHUfVlar2/XCv7TLfjTr5207YVt0PXhYox9tfed5Do8O4GU7PNDa2WB0+93ICliP7quxWY48s6P7RlTLETY+uvedXtpVP4K0G14bQx/MjSnORiNwjk7nOSzA49Mk4Sk/Bb2MPg2j3OUctABtk2ZhX3xVIfL7j3jPVhtHBrz6vsH32TW+z669pVYVy7FxJ6Xb73oEh8CmWybOQ9AJKduwPED1rQQiHzK28h+gqLCrF5HFNsinXentuZAK4Fk5q7WVu21MYsojrPSLRLkSH3CbYZIrC0ld4hNchWERboUBYterS9yLqK11FyBHwE+UiVebbYyvJYtgX+FWL3vlwdjqqEjARbVV3V2SxKnNeydxpzjw4Swwa68yeLVlaH4P/Vml9TDYqCNI7P4eo/S/q2S5xMvgVStUDvho56UJagKoSiw1fo77g/4hl99cecv4lj0AS3r5rrL6Geb0YZlc4jf4+EksQjhDs95B/zagzCN6GIzBjvq4AkKP6FOlAnZL8xwcA/CAcIDd3fMoEoUJ3ut4XC04+/X1DQDc/wHb1cKh +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieve an agent turn by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieve-an-open-ai-response-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/retrieve-an-open-ai-response-by-its-id.api.mdx new file mode 100644 index 0000000..54ffb2e --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieve-an-open-ai-response-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieve-an-open-ai-response-by-its-id +title: "Retrieve an OpenAI response by its ID." +description: "Retrieve an OpenAI response by its ID." +sidebar_label: "Retrieve an OpenAI response by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztXW1v3DYS/ivEfrkE2NhJccVd882XpFdfk9qwXRSHOnC5EneXrd4qSussAv/3mxmSErWSVtLazTkJ+yHZSORwOPPMC8lh9XGWC5WliRJq9vLj7Jvnz/GvUKggl1kh02T2cnaSsLNMJCenF6bp2eJ3ERRHs/ksSJNCJAX24VkWyYBjn+PfFXb8OFPBWsQcfxXbTACplHpCxyxPM5EXUg8b5IIXIrzhhdNWAumVyKFxk52fE/mBFTIWquBxxm7XImHFWjA7EXbLFTMUZ3fzmcjzNG/P6skZ/eLRU/YGW7BQFFxGisllk9xKJCKnebElNACi86HppKFwJqKKXCar1jz0qNiWyRCEKJdbaEZjY0eWLmm8Mhc4C5iu4qsRZH8oY548g+mHfBEJRrNnpjfTTRd2nIo+DMDDUGqBnDuTWfJIiTmA5M9S5jD1l7/q2dUMvQdpyCJCjpooofkh6zIc5hqU+mdZCUKKnC2B7WItVaUIkgKMHQ1Te4fNXGKlEiFRrJWJ5Iz+OugBsBW8qKyDhljyMmo9dMfVhqHVV48+Zzy65VvFrque1zMaviyy0h2e5znfAlVZiJiQlCbibAlCHzSg2g5bXcyc7ub3H6YQHzql1ZTCFSIYWjLDFSIZ0SYTmK2FIs5fk+kXPnW4oUFd8e88dod+ZUbs10DdGXUwBfdmQGK2F/SnSP6dnqJh5qqL0StXPgjMhnDACVmfW/shaBVz8LqOInvUpD3ZPjDUMo7SW5di6/1artZ7G/CySGd3710F0aPdGb8VG7BJgILmTk86RqcEzAcwbyA7ZwFP2EKAnoCv69kcfiAD+AuaXxNlbTvjsEMDdICnfn4Ieqi3ZoN+3pT5CKfkBJyfL95WRkEisPY7CZBGzQdA8rRz+qcuK1NBiQCQSC2WCS90wDWo3P7EY2SL2AQfDpkCygcaOJY8mCN43/MofE/DgAaV5j2R90Sf2hMh56PHx3n250WHeKQRiNB5X9u5uM9R/EmSFpSrqofIEYe4WsLa5iaQesQGX+03zeWhZXMPXBskNGDp0WFrg9Qu0ZYiF0mAuT0Qs0QTCjdDVDEo7SEkk1B8GLEaPU8V2QijDpagnSm7lbCG0evTgyzLxEwrK2eClsVea6v18j30edWnPnxZ81tjjszOCgYXjCoTAYg/IBGRCVa2Z6Y2JjcdQiE4sG4Qtl5MxqBLQUNQJOHNWEW/SUKWWWXvqlllHBHQ1PMcZsfzYvQIl9h68hhG+YOJETbrwPutWKAe0zIPCPejAogTNXpJHYDyWh1N0dlJau5GAL4X7Mj4GKyDBxY58N2Yk3oIfKPmuExEftPvb/va3M2dV12ecwDS+5zuXt85COVDtN2YSlP5ne7O5WAMBiz5hve7vwKJt4wX63aIpKcDQn5ICdZimhQKzonNgxZsTTwOZ0g+5fApx1+ecjRTg3uD0mcgPgN5zBlIb3pwb+T73OT/nJs4acTDxNYvIE/p3845cfZH7g7eTm3us/SydkZ7M83NI/2M9ljv7t4DA3mqHeHwZqfaqkLEe7czQ9ymxGnsbVWqgQZcKQnwRFf9fsTelN0id0HjbJv3oAUGKErV8WrawbIOKUaOQ7uJ7xxOm3unuc4SQsWKlELEhucyLRUzXShCKxvJLD3FTs5Pj9jVWmwZzwVE7YiVgFfI/RJzJL8QAQeBt7uxldzAL3geUz9soDDbu6ZJXM+AhagUkApADIeXQF4qlqQMEgNiUuhO0DvDEAysBSAp8BFMBeBkgP1Rq897HrUXaRqxACawT6cteZcg76Rgur2Nyhi2lOB5sGbIYOXhhvAH/W50vxtixMVh17tGhmHZ35OC7dCYehBDXsxIZgigDZfxi1hc0rDI5Ksu7n+pJVbpgekt4ap+AzW1s/H9yXEBjfJGDOrdN6/9w84BDDgkxImZrSEIKY8ISlO0cz/wYcA5CH0Uqfrg1/lyKv52iehlACgTxhgj0x4F8wKmsgDpdbbthndHnEpK0nH17wXMSPDGtklSxovOkNM6SGnDkiJQ71nUj2L7jNwkq2fDIHylgcRaLlrYVup9mOU8AcUke7YoSuvi8JX9fpoqSHOXoBHnLsELEYkNh0UNow61KRpUa4LsyUIUt0Ik7DlFlhdPCeTjTso7TskbxKc5RgeAfTkyzdywN85pYhbY9JoXxlBaK1eXc6wZK8o8Acwstv0O4ajtmRw0HkLxgFhivelhUaUtoO7Nlr8qriCtAy1QIxqGLpOAVsA2vhxgcZbEQqDBISEdRXi+KmMYcUQg+c/l2U9Mv9s13op8TW5MKDG9OuJI683kIOJSMMf/I5TgoPukAmmXTg7KAx3qPVHZ1bSJ9pONxiLO4MTV8TTLMcz0243l9rElY+9enVe6GVzNBVkbf42HU6FnO2vUPYyB2Rk1DWy6H0A6sSjWadjyBLBK3oj8JuKLMfXDb7FZqyobyWs6bA3RNrLPrTKqQvPRdvKmUZwtl01pmGrzrkLhYdIa7DZPWOZprCN8GWCF07KMHBxNDlp248QxPRvkXTmPs0Tgo9MAdRE3bbVApgJsFWkA+HwC7Z8+NnuMcEWDjkJNW2+gNWHfG+rbstKdV4fYak1CW+wDGwJIciPD6t4CsYrzph8Hr2Z0cVv/7ZFHvJ7Rbo44Z6DM2smhSP6mTCFbxnOwl0LkB3o6JNbKXid4h9f1CyR5u+ZFHfjDVKiJTsHVl5mPY/yAF1zyI2QV/tENZi0tYslNtvvMqyN7t/sKfAOek26/6D4apHXuXuP3cN/XsCML+NEOrxJH2+vtTLfp4pzAuTtJcvE8cSZ3aFV2+6qRv3XiK7/9rRNf6/3oar39rZOvw/f4WyfeEz1uT+RvnfgS0ImLOF8C6m+dDCna13z6mk9/6+Tzr+z01Zz+1olPOb6klMPfOvEZyFeUgfhbJ19sbuJvnfhbJ/7Wib910nvrpHXxYriAw99D+SrvobTvSHxSqPirKf5qir+a4q+m+KspX/7VlJ1rFMPFav6qSm9o8VdVvtKrKvVdkL86T/O3VvytFX9rZappOhdAPoWB+mss/hqL6/f8NRZ/jeUxXGPp343XHPXL0HzvBYRmWCDTZU9sTee8zpngtyiCI1qnIrAhCEfkcihzcM3dmp/j54yYd3bS1gJ3XZ0hbH2v3U/CLVc7Fo2biw3uGd/YUDBqxeJYxOlra1uWVH14DWNxFD3IVFWu/X67WR0fSdLL/JjwXY7ZT3AXtjzOdIJSE+j7do/dSujYOtBFuYXJ2ZZyVeb1cafLtrWDXg+uKY0/Fhs+gjFltv0N8NtZ1gcNtqsceMv+SsA/FZLrwn+sJHcoVwXlDpXrGZ4J2CCpl0E0fQrN4x08XoZIGjUnzVppdpZE21qrDldHehfqMwyXOGc3ZDbmrdZpGYUERZAACPmInaIpnm8h6U/Y5esf5zoFk2hXhcB3v2XbkIMigt8YffNqSGqHRtATFjZj6CEag1EaX9Fy/WPPuLVrZLo7DCBQPoE+53KECdICtnI8x2owFyPEgb1A77wayYsPHInt53nyaXLLvpypOC5nx93oQiLKN030Mc0MfbNoHOKmO/JdmXr9Is1uskle9qcyiAT4cGW9bZXG9fravEyC3XKPYXBdVd1QyRCFV1tGnwmEQcwZqftxNzpUnnSNBDq4a5muDXkUtVATd4yd7xHOdbqkPzw3r52RWdx2pwrujun+LWVLbvcwGTQDCmldN9HDu/mT83VCux+M2+2gTh7ygsCOU/9714cdr2r50EcTZbLhkQwxMsQ8QqzSxw6da773+chjK9Poqwf74erq3OYZ9LHB8dVeemuCGsMCEj1vDilLGcc831rvpr+IeLuW4DiknnQuubnxA+mZfm9jXn2XbMzIujUOHaXJCrdcml9j7HC2ovpMYoIFE8GIWTb8N5Z26cmY3JJsGMwLVh+5FBtwlHhKIxPte2hzeAHo1U7TljqmQUAJXiB2+JpiNjXoTdmaEZ4Df/1RyG7h1WHH7vtQacRFWq7W4MuXaRSlt4pdfP+K/eOfz/9BsQfcPZqKiy+Aej3ev3jILowLqJU5DvnGcr75rttyAvBjgJo1x5OvhFY60DnZWsJKJ9xYowF/x2lpDrNkLLxReaP6zIzqm+/q8WDJz94h0o1lKde0/puWYBMbXOMGQuB2BYVhiKcskrGEbPIckkMsZwL08BUWG0XwMreR6tu+SGW26kCcaEoCxIPAKhPxAeSNMV8YMXjT8qb1GZnWt268Ok1MAfalRns9vjGvkzbkrTBCWPSUgATBY4pKC6y+SFJKjkNrXtWOVcdXvnsJe6PyRvV5GdXzrtFGmBBZCUKNrxTO5mRFa/TW+v/CypW3/58Aiy2TkP2dvj4yy0NzRII7VHoLbeZu76LS4VFVNL8b9+ot3d2BHP3iULUWaLekbZNVefR7/Hw7njjDw5XQq1is/p8db14cg3YTLvFXdR55/NFh+K46NtBTopszs3VRZC+PjyEreLZOFQj2WRTxmD8DhQR/HAVpTJuUSoCYZbGljq+tK/r1Pb5DcBKzRm9vsT+7xP7s0kDU3gzBwbWENi86hKb30lxsc9faXMoFnhmZkxV1nbCd/8B9CErbRRJmKTgrZYupZQ5+ydZ840ocPVm+5AFVZANRnotrXKZHaU622Ca+wOVHBNrL8VBEc0Xnx/oAOAM5guNAmRjYjEbdznZg5bzHUzCQwe2L4yyCLM25JKXB8quWvYYL6uQFYdBABn6/dFEOCkZcYLePHxeQ//2cR3d3+BjLzBAQ8JOc8AK1D/AIpcLfYffhijupJxcG+U/ZJIPpnKPd/E3woJdqPuFf8PMPsd0xXNoHXoNXx20rYFi3OAkCkRVO31a4bJjfv99cgUH+DxsAic8= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieve an OpenAI response by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieves-a-vector-store-file.api.mdx b/versioned_docs/version-v0.2.23/api/retrieves-a-vector-store-file.api.mdx new file mode 100644 index 0000000..d42ab2d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieves-a-vector-store-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieves-a-vector-store-file +title: "Retrieves a vector store file." +description: "Retrieves a vector store file." +sidebar_label: "Retrieves a vector store file." +hide_title: true +hide_table_of_contents: true +api: eJztWl1v2zYU/SuEnjogid2i3da8ZV23BWubIkkHDE1h0BJtsaVIlaSSeIH/+84lJVuy5MTZ+lLABYo4Enk/z7m8l85dYoUrjXbCJcd3ybPxmH5kwqVWll4anRwnJ+wvkXpjL/Bf/CaVOJt+xgNmRYnNQnup58zngs3w7ig5SFKjPR6TJF6WSqacJI0+OxJ3l7g0FwWnT35RCigwQR42ltaUwnoZjZFZa43zFmqwpmvbBy2/VoLJjMyYSWHZzNiVMcnyoBE+KGnGK4VXyXXwb+LIwaOwcVNP7TKJaCk7YFzd8IVjV30RVwlp5x7appWPHm36y7NMkgKu3nc8N1qczZLjj6sdulIK4la/T41Rguv2I10VU2HbT2pHW0+4tXzRflBbsvy03HT4T7E4vOYKsV27wLhzJpXci4zdSJ934pzmlf4CfQiBxYr5og+ki/oNqxwkUKIc0OE7+GFSe8OCMIcI9SKxBSvxbT/HgKKjDPPKm07KVw8GDSQxreTS4qskWEwfC+A5ZY3DyXK5LZMzrpw4AMW+VtIKwPljNPDTQeKlV2Rpi1qvaoGNFSdDJp709LMm4MG+iEMWcBgC6o4GEv7Y8DlPKjsBbD3aJYRxeR3E+EsrggeNuB5mXhk9k/PKhhLCSm55IbywbsXzDVmraMCy+30OOybmWljFy4k3X4RukxQ4FHMQquXz8/F40913gXXMzFgUgB+sFsmmwt8IoRnPPvMUJaNBNbwt+O0k6nfyH7GT8p/7yt/yW1lUBdObRsDLqOyAFZXzsGRlzNPxmHGdsefjlz8+DrmD8drmym4Ivwi5iyn+DzRaoeYx2gYQu4VOocSlDf6mqE1r+IFUUJpJElRIzaGP0lcjbPEOyyC6trLAKUgwp/OQOP3gybevZvdWs1ax2Je1fVnbl7VvWdbI3h20hpbPCuoFJ9xvyXHbnktZCPhVlOwmR8JW/d4NR1+ZZWgIgTF62uY8aVHc+YmwNlbYrtAnZ2WM6w/sNa1A9wgCFZFVchY1gB+pcI6CMON4kD3MIZOJwT68X0GEBXJq8/q993ohRW2icFbAl9tUCDgcj5COP9EJUg9PsjA41b1xmDxABfKgioFBPB2fDxa4rtQ/qoLrQyQrC+kOxrJ6N4tLp6sevJH/KA5RvNYGDeOWJsc3yOXrOlYB5ZUbqI2VtVRXWnmLS0MpaEaOgdmgF/TUFKUCsLN7UyP1BKrmmGXdvetSrlOh1APSaogttwfhIvqNCFQUrsl00R0PB6pkr0aSMMqeK1GD4zw1xeGUS9dMUSyKhZb2aDrZZag+/bWJdOf8pbGeSx2RUit6HE4kUW89/a5n46HxsVNfVmjpBq3v3PawnzV6Nyb7UuiT0/qKg4XVjJazaGeoiPDx+dDNyCVCRP6hsIU6JjVGZpkxohdXVIlCsflG9yFrvjxQav+4vHzfUCZQc13SH0p9LEBhMboc5nJjPXNVUXC7aEAR68dNLtOcyei0lRyEDR2lrt8HPeFaAaBRu2qOq0m1Mhq+sbxbu1p7OuaQJgkCEkUf1tU6Nk7Yh/M3tTMgOPUUgU04jKzAbnEtWEGYaJ8sfGoqH9u0UqRyhnPWpGkoW6nYsOsx/FhhPGZrFbwWqGP1HA5ec5XHZtYUwYaT96dH7NxU81xRQ62UuXHs/LdX7Kefxz+FVlrccqqRbXyFfrDR9wvP2HmE+NqeHZFfM+fZy2HmpEpSmc+xnW4REXKDzXrRCCapAMJcXlO3WZgKayi2aCT2pNqT6jsj1bOXa32XQPpbQnrNLNem1t+mAicQoaZPDErpUGShfzxi75XgMMgDPXyOE5kpvLTNSfVi20kVe1WGcBKVBMJDwKq0uEW86U5X1GHYU2tPre+IWi/a59UpIRuWs4uI9rX+ml4nfcg3wciO2FkFJAhehFNpSjcL2oTverKGXquevPc12T2C96Tak+r7ItV4SNsOFAosIajxuSNv4lx1epb07jrO68g65Ll3z0tfI6+vpcKQr+Od+ubARynH45L7vH/nBO93GWbrq6hWukn/OineVkhSj6LNF6x0HVBbR3IeZ9X/0/yJbl18brA0mYtQLkjlcTK6fjoCzDSX9KkdNDe624jhchSu1kd3tfVLGrdD+YyRryxlPfe+PB6N0LYc5sYh84dK8YIfAjHpl6PUFOHOwwngQPpF2PhrUys/fqJ3xJ7gRA2sN7QfIzf2s4uaQ4FbYa63Lobr+ulABKnkuA75eLsctCX7nIe7pGsJIVeabfxDfRNhrhA6Kw2qqQsXyZAi6SbORj5lBBQqtXbGU+GiUG7FFaFIGRuKRV/4lOYjhayGy5po1VsUYxW/xygRR1Q2ikmNnwdJ0QnE3fpUeXhnDR0vbv2oVMB+uIGyoSpH0HyMsY6woRw83bhgoap03CdgQA+9atCPXBNESOLd3RS96gerlkt6jHbXEjbwMRwYUwICkJJJR5+zVUXc6ueT85ocP7D/R+/BeDR/KKHp/in8/QN+w8cvYjFQe4j438Dyb2RdE/3w1xw5Dkt0BRTb+PYkTUXpW/t6XUinmPz++hLl5V86p8Bc +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieves a vector store file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieves-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/retrieves-a-vector-store.api.mdx new file mode 100644 index 0000000..167de50 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieves-a-vector-store.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieves-a-vector-store +title: "Retrieves a vector store." +description: "Retrieves a vector store." +sidebar_label: "Retrieves a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztWNtu2zgQ/RVCTy2Q2GmxRbd5y/ayG6BtgiRdYNEEBi2NLbaUqJKUE8Pwv+8ZUrJlW7kU2YctkAJFJJlzO3NmhuQiseQqUzpyyeEieXlwwH8ycqlVlVemTA6TI/E3pd7Yc/ynk/E3vAhLFQSp9KqcCp+TmIUlwvGaQbKXpKb0+Jm1yarSKpWsbfjNscpF4tKcCslPfl4RjJigF4KVNRVZr6JDKuuscd7CHNZs+velVD9qEipjdyaKrJjAk22nkuVea6RX40TWGj8lUWYUZbZNNdGzdMfenpD6Ws6duNyQvkzYZmpJespGsmtXAZwp2R39F6og52VRieucyp0YxLV0olHIuktZ0P34PDsJT1I/F58hIMykF5zaySmNxnMfke/xtIHoYNsAEwOywlUyJVE7ysR4vuu8KkXUDmMTpWmUmrr0bpdwH/CjAA9Sco75BUR8jcDD8v7k7t3Do9QUlSaG7f4kfK6LMUgEmNhLB2PSi1zOSIwJSXF1yo5Naq3nrZcxHaksU9L6MUYyDk3ZbvRrrYybVI/QH6WFN61+VqnKEd6mll9/Xm9aW4s6ABRjYm83APHGS/0Q2vM6UW6pVrsFkCyhVWaZioQ+7eR4IrWjPbSzH7WyDNHXTtK7uVmBuBl66+0VHpTX7G6n7TEl30a6woNIyLv7SNf4ZrRvI2QtrW8pR7qpEIcbyYkHYj2Nsh+GRWJKOpkg/JVECaZC4+p9bIwmWXY/Rei7X5qIOl+ktXLe/dB4srxa3tFx3nMYofeLymAMzG9tzquAH9IoOxbu7ZlKaxGVsxktnR/J1KsZPcIUksaKRFCk/FyYPrLuJQV5mUnfO+n+xwk8J88hfqf5/kxqDNdKKts0EVQSal1I7yWGeOgmjytTxSWyQqUzLzdn0ubQWBVhB+P+0j1pNW+N8orKo+NmayPCShGdGMB9DuC3vs3QBUJl58GDMI1VCYBUJqCjkBrcLmK7+W+2PzuN5jaS/nVxcbqek1ngXoPFfduD99ZyRfJibGSEy41Fb6qLQtp525worLnOVZoLFYO2SqKHcTGDD/H3YCdwyaPDPtRyXM2mtSkRm8jrQpb7oEEmx9gIdGQ23ImzCzGjsf/UJuhIfDn72ATTkDnsWUBkS5AmTPoiblg4n7F5ybGpfTDuKkqx7UuFSePwS2nLr58h/4rFMVsr8DpcDjDdAl67excTa4rgw9Hp8UCcmXqaa+61WptrJ84+vBWvfz94PYiNVvJo6vILVF/b+wO7kLNI8bU/D2R+Uzkv3/RXTqoVT74c4nx4AOQGwuW8VRyGvhRTNGf8LbjUA7bou09F9VRUv1hRvXyztncBpn9ipjeV5bql9Y+p4wGDblKijMIpQGDbREKrQvmBOMXAh0Me7JFTiSLR+NG2k+rVbZPKkZ0h+YCTS4kADxOrLrEdQmXgjRoYnkrrqbR+odJ61Z1Xx8xseC7OI9vX9pvyOtqlfAtGNhAnNZhAsghTKZzwSxPudbK2vFaHup2bsTsUPxXVU1H9WkV10GftASUUqoSpJqeOo4lHquMTdn/T3bMGWYc8b18YV9JKnOTIunD0jVebG/epIxXvbPC5kj7fvUVC4Mfv+u5TulllW2vsva2Ri51KbA/Py6twvMwNliZTCrXJpg+T4ezFEDktpeKnrptuuNjyesmn1dCbYmy1ZUhz76vD4RB7gv3cOMC6r7Us5D7SkX4fpKZI2LojgKz8PAi+axvR1yv+jakZnG6y9pHlcZSFvDhvCBqIC/tsPMI0e9GDHNez22C27NZaV3O4BECHmikouSzF1j80j3h5QGVWGcU3trJsrzVTYyNZM74r5D5mJzJtryelpcuSCWdsqMRd5WM+fGhkMdw3R68+odNpF0hdAUe0Dcak4c9djNvAYLHu1ncKNQTxdOOHlcZOLFyd29DoIjW+RoQjORj5FwH+DkHwfrhNbKSTWcDii8UYe70vVi+X/BnbRcvpx2NouGPONciQKcfP2aqj3BrPs7OG78/Fw2ukN9D2Jqmcc0x8M4Q3PH6neU+xhoulHC2cLzDhcVx1lKZU+Y78zmzcqLo/31+gDv8FjzNcNQ== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieves a vector store. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/retrieves-the-contents-of-a-vector-store-file.api.mdx b/versioned_docs/version-v0.2.23/api/retrieves-the-contents-of-a-vector-store-file.api.mdx new file mode 100644 index 0000000..f098567 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/retrieves-the-contents-of-a-vector-store-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: retrieves-the-contents-of-a-vector-store-file +title: "Retrieves the contents of a vector store file." +description: "Retrieves the contents of a vector store file." +sidebar_label: "Retrieves the contents of a vector store file." +hide_title: true +hide_table_of_contents: true +api: eJztWGtv2zYU/SuEPq1AEqfFiq7+lvWxBWuXIEkHDElR0NK1xZYiVZJyYhj+7zuXlGLJdtYE7ZcCKZBaD/I+zj3niuQyc+Rrazz5bLzMnh0e8k9BPneqDsqabJwdCa18EHYqjk0gp0nOqXhlcW2CcFTDAq6UmYlQkpgqTSJPb/1Btpe112xX1rVWuWS7o8+ejS8zn5dUSb4Ki5rgzk4+Ux4wsXa2JhdUCo3tflJFb6APDk4xcBjuB6O+NiRUwUFNFTkxte42tGy1F00ZWdG3bf2NUZx4f7YMGDtpQgprM2hZFIonS306CN8aOplm48vbGabRGuZu7yfWAlnTf2SaakKu/6QNs/dEOicX/QdtJKuPq81k/qLF/lxqYLNOQUjvba5koEJcq1AOMu1VbuBtL1OBqp3pD2uW3m6DDMMeZrNAN2EL845ZPGlP5I1zuNMLYQ3+u4pzrjKhvPBNXVuHyDnWaOqb9bxAdjIPjdSCJ3Q8zVaruyo3ldrTHlTytVEOrsaXyUXr8SN+VdDs8h8gYN05/qhN4c7UGD4xdbYSUszjNOF5XhIP35F0eQlx+UaHg2yrlO9aQeY9gz5ZXBfwITl16uqJY8D0NRl2Z/wWs9rs/FnbUbbS716kQB3BOs27vtG1DE5rByoAYcUZ/bqrQ3FZORsCKtfSC2XAc1UwkpXUkH9FxY/rRD7I0PTZr2B2BqFu5vvnxcWpSKORXhE11WL3LaK+cY6bFg/eAxy+BNFB+KqSbtE1JIpjrksFpqiUtFMSbOB+J037PvqJBApS6ft6TqPZtbYGuYmyqaTZdyQLOQFHe3MG4bAnBXFLk98jy19O6sTOJ+JIfDh71yaTI/gJicajKQXbEYVExWRQhusZKyfkxDYhOvc15Wj2ubB56hg5bcT1EDW0Fe6qdQtej/sRpjvAcwOecwxHp8cH4sw2sxI9bGq1ttdenL19JV78dvgi6ptuZFUnZnT8AtXX/n6XhThLFF/Hc0/mt8p59nK3cnKtuIeUmO5j47UWk82iM8xWQYSZmhN+K9uY2HuCik3iUVSPovqZRPXs5drfBZj+npneKsv3pfWvbaAJIEQ3OVHBsLHYsFTCerhS4UCcYsGGgALYI2cSItF46bov1fO7vlSe3BzFB5wsJQI8TKzG0A3w5oUYtTA8SutRWj+RtJ73v1dxq4jIxXli+9p/K6+jbcp3YBQH4qQBE0hW8as0IXx4jI27uaKTV0FTicXxju3q/xh+FNWjqH4uUR3u8nYPCUWVMNXkzHM2aa92fMLhb+7KErL+ftswCEE67A8hbx/PM9JJSpaGfopD005SsfVahnLnNvz4dVebgRP2j0/p4DSpV372vy5ScA2KtiXZ7pSET0Xa6Nb72/tG9X2egTIgKi2GZjOK7YNdjrPR/OkItDNS8VUfND9abmC4GnEQeN5GvxrltwcLaRGRKtA4ZkMZQj0ejbCc2S+tByP2tZaV3AeT8i8Hua0yjsoT+KHCIk583fXQy4/8jlUVk2kJ947ni3OeL85bbUXNwT87T7DNn+5AkluRH4hS9ttE33IoZRBornMFI1dGbPxD36O43yBT1FYxNaWJazHlwBaXdFYwYbgFu6nMI5VhVDq6YjZp62IT2TY+4X2TRnWdnHVRvUeT1j7qsQaO6HiMScujB4tlAMxy/fV5uKWWYnzsNKo1NMIRxtIvW3JdplokenGNnsZC9SiG+/G2UCPL+NVaJb3THiYT214uJ1jtfnB6teLHWDA7ZhEu4ydnwpQBpwrl+bq47al3IvDLWSunJ+L7GsJOZLoTS8PnlfHYE3e4/EKLHd2KW8UPiPwHRdfVIR7ilvjcYl3B2Ka3R3lOdejN21rHDNrPH28u0JD+A+kLXhw= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Retrieves the contents of a vector store file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/returns-a-list-of-files-that-belong-to-the-users-organization.api.mdx b/versioned_docs/version-v0.2.23/api/returns-a-list-of-files-that-belong-to-the-users-organization.api.mdx new file mode 100644 index 0000000..17fe708 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/returns-a-list-of-files-that-belong-to-the-users-organization.api.mdx @@ -0,0 +1,68 @@ +--- +id: returns-a-list-of-files-that-belong-to-the-users-organization +title: "Returns a list of files that belong to the user's organization." +description: "Returns a list of files that belong to the user's organization." +sidebar_label: "Returns a list of files that belong to the user's organization." +hide_title: true +hide_table_of_contents: true +api: eJztWV1v2zYU/SuEXtYCruMUK7YG2EO2tluAFAmSdMPQFAkt0TZbitRIyqlr5L/vXJKypNpJnK7bUCAFikoWeT/PvYe8XWZWuMpoJ1y2t8yejkb0TyFcbmXlpdHZXrav2aF0/qgSev/glVTiJG1hudGeSy31lPmZYAqrmJmwCda4YTbI6LvQnkTyqlIy5yRy570jucvM5TNRcnryi0pAkxm/F7nHxsqaSlgvo1UF991V3Fq+wCLpRenu3p1+btc5b2FxNM/hQ0b24rUQE16r3g/dMJzBwyiLkaABu5rJfMakY1xd8YVj52HfeZZdw7Zik8J1gbSDyQJBkhMpbCM055qNBbNiIqzQuSiY1CHE+8cHTOiiMlJ7R4rGCy+6QcDvYirsRmVOfhKUH58UD0hqFABJuRXci+KC+y3FvdHyI/OyFM7zsmKPIMwJxLRwj9nEWLgi9EoXu+KOJRWkTXysJKD3b2lL4kkTvWteiu0SQiu7MSIJVW0r4zYKELous723GXcO4OeUFOSE+3yWvdsknjzUBfKZZPZVQRcvCknruTruoHjClRMD1OpfNdwqSOMK7YBaA4NeEnsx7oSh9QcWeukVOdQW91Ejt298XMBoRVMFSChKRuoWnZ1FjqA6pOj15Rx2ekQSFLI04+6iNLYb5LExSnC9ZsofMwFlljRawTj+0sbYdRifc6n4GMLHYgF0YBUqtOJTEbFgnb/YpjgPXrSpwZ5UqLrtc4Q5iJU69DQSrvi9ZdOWbUTf3cNoY6+HrX64Zw+jfdTD7oPF0KA7OewEug3LyosO7DbzyprZK8KhyJCBxDgx34jbBtRdk/nfbyIzigCZjjYSWpLUc65kwSC45AryS1F8PeZCS/D1Nu35t7OzYxZXg1WLANYUpLvQ9NJaGB8WDxhnbmasZ64uS24XDc5EWLNKNjltJbpVCCi4Jn4PekLFgtXVtprjalKtjIZvbFaXXD9BIypCHXb29MwJNKmpaeZbePnoqIpQfMz22ZuTwz5R1g5NyBtkFrvFPHUEqSmfIXOMj03tg3JXiRxkmzOT57UN9PqZXfeBfspwk61V8DogD2G6IXh2hWxryobhh+zE1NOZWiA9Spkrx05e/cJ++HH0wzBSJwgwIqPBF6De6vuZF+wkQry1Z0vkp8p5+nxz5eRKoioYCh3kiwdvDDbrRSM41CNnUzkHGfPS1Do0e6Lth6J6KKpvrKiePm/1nQHprwnpqbJct7T+NDVqYk7HzlwIOt6RUouTGNiqlH7IjnGSgUEe6OFTXJnA/F7Yhqme3cRUTtg5ko9wUinhuFMQsGqNkx0qA28iheGhtB5K6xsqrWddvjogZMNydhrR3upP5bW/DvkmGMWQHdVAguBlYKWxAPFoE27TRVNeq0PxhsnGjYIfiuqhqL6tohpt0rZFCYUqIajxqSNvwkVqfXxxInxttaMUd+d8sJd7hJ/yTtEn85EJ+53DyXLKtfwUYk/jwIpbXgoUO6lZZnEok/GJD7UgSQm41S7WwrTPYKmDyYRUCKdTZntBHrLLIOMy3GN1c8HFLTuYFmcUji0MGkWleN67bA/ZK8hsIDhgckILcaj9IBpPm0Mz1wWec4HDLdsdjZrpxYAGcnQhvZJ+Rj9eTIwZRHWuHjvajSLLuVIBqVLnqi4gnWz+Ka0nk4wtaKph2ER4oJosRMp8GF6sBgZkctZFX4LjWjNKJYTEriIdziJ3RTosYiZGSNflGDZBeXKWzAuDScICdd/DsJz8shyFj4/+inrwbgjX7mgwQqsPzyQvdWJKVPy0nS9Nw+w6E6J1uzOn1ARjVMeLoP+ynY9ddoaJKbjJR8KTyy9jW3R5ym6UQ45ckpb4mZ663++Vmu70ME/G92ZyycMbnerMiUSx6iJu2ItTM+y7NVJHWi1SUlNVBzBTUOJdLkn5Yv82TUfXZ4/HK1v75v0eLqrNyBTXVtRj8H7T6AeS0WRmBgZmUxHomPsZXnbmuzto45pLegpu4mM8YseOVFvqlTPvq72dHRz2n8yMQ2ifKMVL/gQe5B+GuSkz0uEEepL0i7DxRXPCePuOvhHnhHg0Yy7az05pPztNzBNne4OMlEcv57sb5nVE1K5HWbxLol3Jodvh6DGXEHKu2Wd/cCoQoW2v/v+gKUtpcdywET8By1RvdoJOmbo7t+Kc/qdHGRsodl34mBqkAulaalbRqtc4wigX2Ap58zgPUEwSLP85m/QitWwPa19BdMKxR/fdAWPIMIMN4FgmML2N2YpwoizuphF7IE6CDa1ZLse49b2x6vqafo61R3gppKNjUVtDNzrzQH2fU9/G5HwQi85RInQIvGfUB+8T6/+e/G7xpqHrL/Tm/2G/W/xpGPsL/bkPR91iRcuHrR3v6AV3oTF1a7RwVDAuLjCWajVu289zUXWzsXYj7BHPry/PQEV/A9OG558= +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Returns a list of files that belong to the user's organization. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/returns-a-list-of-vector-stores.api.mdx b/versioned_docs/version-v0.2.23/api/returns-a-list-of-vector-stores.api.mdx new file mode 100644 index 0000000..1bb923d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/returns-a-list-of-vector-stores.api.mdx @@ -0,0 +1,68 @@ +--- +id: returns-a-list-of-vector-stores +title: "Returns a list of vector stores." +description: "Returns a list of vector stores." +sidebar_label: "Returns a list of vector stores." +hide_title: true +hide_table_of_contents: true +api: eJztWW1v2zYQ/iuEPm1A3rGia75lfdkCpE2RpBuGJHBo6WyxpUiNpJwKgf/77kjKpi0ncZq1Q4AUCCpZ5PFenufudLrJDNhaKws227/J9nZ26L8CbG5E7YRW2X52wP6E3Glzin9wJKw7iVtYrpXjQgk1Zq4EJvEZ0yM28cuZpfV2K9vIaB0oR6J5XUuRcxK9/dmS/JvM5iVUnK5cWwOeqIefUQRurI2uwTgRtIs/z9dZZ/BsXFfAiDcSH2Wkg/8hteDYb2S0i4kCNREjAWaDcXnNW8su/K6LLJviRu5STbgxvEV5wkFl79dQFKu1S5X5pMQ/TaoHG6GzyH+p30iXdewNewZhz4PtTncH+3MD3EEx4Om5AsM3BtOTfyYqsI5XNbsuQfVsYNfcsiiQZCtewf3++enYX3H5M/uAGwhQq5zTWD6GwbB1YG/RNLpoZ/kAwjHuZbbmObDGQsGGbV95oViQjoeNhIRBrhvlbJ8f7/AhQxzkYC1RAT3iGjTcL18d3I17cJTrqpZAbrs/CB+aaoggQjeRlhYP446VfAJsCBgU2+Sk2KiRsu20DOHIucpBysccUpBpwqTWz6WS37h4hPywmzndySeRQg3wbmzo9uFy88YY5AG6Ygik7YJDnHZcrgN7WsfUkmjRJ0A2Ram8KEQA9MckxiMuLWxg9v2nEYZcdJ4EPY3NzImLpnfaXuKFcJLUTbI0QfJ1gCtqEAB5dx5JD1+09nVwWQfrW+gIX2u0ww74yKHHViTK1W7ApK7geITmz3YoRCpKnN0PtZbAVfpTcH36S7Qo+SVk7uSHqMn0cnpHxnlLZvjqxGqNhaq9NTnPDF4nUSYn3JszhZQsCKdjJLduwHMnJvCIozBoJIh5QcK1TK8C60ZWgeNLFfApBPAUfN/xBdrNCZdYXGsuTEwiyCTkOuPOcWwzfDZ5HE0FUWTmlaReLtakxaIxI2Hi49XUPe4kL5XyGtTBYezEmF/JghJbWc8fRysasbg6ljODqFqnWUnQdPim477f3iuWsw6QGFPzsVCeRzMQf/NxHrkPOK3kdlBRWPd7EEwyXozs4vl/lYCCPeHxGI5/JGixoWV8ggmZDyUV2FYrqoHCkg4PBNIMQx4Nid6rgZF23j3HzVrykdGVdwxVt6VGfDol/X5Z1eOfoTtJN8wXvmsTCokkCoa7Ky7Rx1UoS/9NI98rSLclsz/Ozj7O+6nC56jomvug9NYYyty0GBteZkttsIY1VcVN2yEL/JrrUuQlE8FoI7gKoMK8EZ77czzH8GVHrntyWE1HS63QNlY2FVebmC4Kj51kz4I6ocdBm7EBeBBhDtink6NoTEx6vrfFhGcAd8MkolkoimcocnyoG+cPtzXk+HqQM52HJimHJb0egu1ZtgvRmjkvgbZ30y3OMwtwJh0OPh5usRPdjEtJNVlKfW3ZybvX7OWvOy+3QkHm1MKk+EKoz8/7DbvVkwDxuT5rIj8yZ+/VaubkUlCHhARmli6c1rhZtZ1g3xxyNsYijv9XVBK8b7E+P5PqmVRPjFR7r+bnnSHS3xPSI7NsSq2/dRNeROFrDlCAf1tk2F5T6a6E22IfsSqjQg7Rw8ccSSLxoekq1YvbKpUFM8HgozuJSliqCwJWo7BtRmbgHUQ3PFPrmVpPiFov0np1SMhGzdlpQPv8/Eivgz7kO2cUW+y4QSQAr3xV8pMgpf38r+joNWuFewPfOwQ/k+qZVE+LVDurTluDQp4lBDU+tmRNeBU7PCb1l1++XGOUpSjf9gWk5objiz8Y6yclYRKe+SroB3t4g/XTtD1XHIRS2Q1s5kPH+DpPkR1SB0sqEO+P/HKKveEIOXzoron9u4jlgu1ihvEXJCwmAEL93g5pOQ9ejGaPyx1HpzSXiVZoU3jW3mHFKfHOr+tG7VfzwcmVb4Rngyp6Go3bYlfc5leBiTYHVdBbbZBDRlzRKeExXaXP1zOnmzol1oQB5j0xQYRYHT7cIL2ou5+PIEhpknFFfsUwBFtonuHnUeh0ocCyVmOCriV9gkjGGd+s9hBG4cPCY/QOQn6E4pd+DlZqXJuNwRcH7kq82Z7sbmNSUVzQVfp9yg/RfCkMHGoMMbh0rt7f3sYWdLPUFgG1KSWv+CbCKf+ylesqo7MsoOXCtX7jm67unV/SM8qEXsWYJI5oPzul/ew05sMwW9rI6PDgy8lu/7MAjYGEXUikPE3tqWTvUiyIE4FCLhRb+oe1Ksw0EdO1FvQhKbJWGCyCJuRGD3eipBlhOOLAkxu4oM+yUhuf+PvCh/SuK7EU+M9gQav3WFil9Tm0Rj9ilSKfRHCtkeAWXHEz7xHW2RvB4eCr20ZgCT/F8+G9ibA4D/4OwKA47PpgpODASBIAaO3NzRDfKj4ZOZ3Sz4EMFPlCWCq7xerpX6r1D027K+3/Am1SIfxYG+/9PHx9M/6fvHuHPV2t+EZ7vnPmvUPxrix8L8UfmXrv0HxWGeaqX9INNrlDSnjn9D2lxI6UvtshScKugzyHOkVer9VfyOG/vz3DrP4v+XkD9g== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Returns a list of vector stores. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/returns-information-about-a-specific-file.api.mdx b/versioned_docs/version-v0.2.23/api/returns-information-about-a-specific-file.api.mdx new file mode 100644 index 0000000..351153b --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/returns-information-about-a-specific-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: returns-information-about-a-specific-file +title: "Returns information about a specific file." +description: "Returns information about a specific file." +sidebar_label: "Returns information about a specific file." +hide_title: true +hide_table_of_contents: true +api: eJztWFFP4zgQ/itWnkCCFtCt9rZv3O3uHdKuQAUeToBWbjJtvOvYOdspdKv+9/vGSdqUFl2R7h6QygtpPJ7PM/N9YzvzxJEvrfHkk8E8OTs54X8Z+dSpMihrkkFybsRlSeb84rPSdDn6TmkQqTVBKqPMRIzxVigztq6QPKOXHCU8TCawL1mWWqVxpP/ds8N54tOcCslPYVYSIGz0iomlsyW5oOrlNK9Xdj44QNYAHgMJg+NnRmNZ6bUX3QhuchK1L8GOjsRjrtJcKC+kfpQzL+7jvPskWRwlKtsGuOmwjjtDmGqsyLVOU2nEiISjMTkyKWXIjQiwP7+6EGSy0ioTPAONZqEOs8HCe5qQ2wrm1U+EMI6OGPiIvdYO4Cl1JANl32TY0d2tUU8iqIJ8kEUpDuDME3Ka+UOBQiIUMkss8Si9aCAYjZ5KBdb8X2iNe0bi30YWtFtB2LKbI/ZQVq60fqsDMlWRDO4S6b3CwrgoqIkMaZ48bHPPEZoM9Wx8rkMBS2aZYnuprzosHkvt6Qgy+7tCWBkjLtkOqrU0WCviWo47aVjFgxUGFTQH9FybG7mpDQRbtCpAQSEZZVbs7Bh5pmoPIXFQv2zrCJwODggFjeRQZiq1ygRqWUjNnYCy/64LoDih2kUof97cXInaGg0qiwRosvRv/PnkHBYfjY+EFD63LghfFYV0s7bQFG2WrYODdgq8iRyG6uvxiLNg/2iQelfk2pqhtTWITeRVIc0xKJHJEerWmbO2nNiwDNM33SHKg8uyJuihOBe3wy/rLavyoEOwqCxm05REYd1aZxdyZKsQwX1JKdpeKmyaVi42umfreo0gmgq31Vomr8PymKYXktduYWLsbNH22p4Y2mqS6xnKo7V99GL4+Xfx/teT9726iaEV1cxo+QWqr/B+k5kY1hRfrWdH5jfKOfuwXTmpVlCFyDHd80OwFpPNrHXMXkGEiZqiLcrCVrDh3KrYAfai2ovqLYnq7MMK7wZM/8pMb5Tlu9L6y1bQxJQPACkRb7QM6rAnCq0KFXriSpPEggLYIyc4fQqNQdfuVO9e2qk8uSmKj3SylHAqy5hYlcEeC2XgFzVp2EtrL603JK133f3qgpmNlYvrmu0r/EZe55uUb5OR9cRlBSaQLOKuNCJsPMbGe03Wymt5xdpyPXzR8V5Ue1G9LVGdbEPbQUJRJUw1OfEcTbxIbV4khxQqZ/yWhMpVOvm+x19RSulw64Ou2eM8qW/C8Tb4Ld4cFXssZci3XocvPnZvqFzTirPD7MzB3OawyTirKgRXoSobmmyYtFggHiwotzBNJhRVyvCDpD897aO6Rip+YkDfnzcrXcCs3oPrOCrHycxDKAf9Pk4Dx7n1SOix1rKQxyhE+qOX2iJhNE9IrwqzOPFj24LuHniMcxgX2dTrC88X1zxfXDe5jBkGPoPXqZmebskWK9mvcVp2Vdb1HHIZBHrTVMHJvRHP/lBGisf15acedIV4lFEO/cjVNM34wxl3MDeWKfnaqXR0z1/VtHVRg5vOR3zt0FClk5N2VV/R47SPdC6RRzQMzknDlVfRbS0p81Xjfp2XhjKBnkK/1Dil8cpiyecNWe7qGtR04dqcNt84WPyDlt4oMPOCzefzEc59t04vFvwavHVMCDzG5jvi6oMemfL8nC27y4sBHQwbxh+K1ylla3DNSzCZg5G64l94/EGzjlwXDzDN0cqxZ/Fq69HzNKUydOZt7JFrmvvj0w1U+A9q141I +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Returns information about a specific file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/returns-the-contents-of-the-specified-file.api.mdx b/versioned_docs/version-v0.2.23/api/returns-the-contents-of-the-specified-file.api.mdx new file mode 100644 index 0000000..e4ef53a --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/returns-the-contents-of-the-specified-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: returns-the-contents-of-the-specified-file +title: "Returns the contents of the specified file." +description: "Returns the contents of the specified file." +sidebar_label: "Returns the contents of the specified file." +hide_title: true +hide_table_of_contents: true +api: eJztV11v2zYU/SuEnlogsdNgRVe/Zf3YArRo4LgPQxoMNHVtsaVIjaScGIb/+86lqNiukyLBngLYL6ZE3q9zzyGpVeEpNM4GCsVoVZyenPBfSUF53UTtbDEqJhUJL2/ETBsSytlINgoZhBRTbaVfit7FoDgq8jx7kU1jtJLsZfg9sKtVEVRFteRRXDYE5276nVSEYdTR8Itxdlas8TsqfnswI/q3pRDFDTLRdiGNLoXzopZm5nxN5f/JpfGuIR91B0qIMrZha52G2zl5LNxN66/J5EJ0q4FTiRLuyrqzDdFrO98z/eA9kk+LjwBsqJyPIrR1zfi6mYiomNKam0qrSuiuaK8leoGChbR5PsVZs/8otXls5G41hzbOojZRtbW0x55kKado+5bNTjocSVvUbNUjqnzxJY2keSnOxNfxp1yMQvJTEm2gUkSHzsKaFiRq5wllcj9T54Scujam4KEhpWdaCadU6z0h/E95ITFZlrqLd7HV0Jk0gY4K5o/24Mnoqu9w36078K43tEwwPQBez38x865OOZxdnA/E2LXzyizRHmPcTRDjj+/Em99P3gwYNLqVddMxo+cXqL6J94csxbij+CafRzI/K+f07f3KUUazgiuYBx5E52Bsl71j9goizPWC8F+7FmsYW13TQVQHUT0zUZ2+3cSbgOmfmelZWWFbWn+7FpoAQnSriEqGLR18kYTRtY4DcWFIIqEI9si5hEgMJv0g6+31QydVIL9A8wEnS4kADxOrtXQLvCOeKMNwkNZBWs9IWq+3z6tzZjYyF5cd2zfxs7zO9infg1EOxJcWTCBZp1NpSjh4rIsADHNZXiXNZGvivsR+5fggqoOonpeoTu6L9ggJJZUw1eQ8cDUf8bEWOPfdXMcUW29Dyi8rI/SYZTjhmr/0+GOukV7WBGGzy1Vh8QAfPPuPZm1pdtnIWO1hwufe+fvec/pyRFNbhofpWYG6+bbJcTZtiL5FW/ZEmam0XqMgJFQ5LC3mlGTK4UfFcPFqiPZaqXnEAcNwlTNdD/tNAK7T9tTV03pGtYqxGQ2HuBYcVy4A2WNjZC2P0RH1Y6BcXXDUQMBZx2UyfN/vRVfXPMfsTMnmxn1ie3HJ9uIyczRxF/E5eAfR4tU9qLGkww655bbctj3HSkaBTWqh4eSbFT/9sH9QureTLRunuc/SpjuN9mi97/haAlbBW5mfSUWhcyo9fbPMOeeTGPedT/n7w0CeXs77rD5jszMh8boBjtg5GJPMmafxbgeV1WYLf6KbTJ5It3HYGFzYOLfU9FWmzVXXhY443B1+SuTB/2hD9D4DNJs5woar1RSXwa/erNf8Glz2TA4M0448ZSaAKqUOPC7vtpwHa3sxzip4KZ6mnnvLzC/Bai5LmpafMPxByy0Jr6+xtML+joOMs+1mz5SiJm7Z7R2cOzr888MEyvwPwgJeUA== +sidebar_class_name: "get api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Returns the contents of the specified file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/run-a-shield.api.mdx b/versioned_docs/version-v0.2.23/api/run-a-shield.api.mdx new file mode 100644 index 0000000..0a631c7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/run-a-shield.api.mdx @@ -0,0 +1,68 @@ +--- +id: run-a-shield +title: "Run a shield." +description: "Run a shield." +sidebar_label: "Run a shield." +hide_title: true +hide_table_of_contents: true +api: eJztXVFv2zgS/iuEX64FEqdb7N7e5p7Sbu+uQLstknSBQxtkaYm22UqilqScCIH/+82QlEzLcpw0PdTRTrHb2hI1wxnOfDMc0eTNSAtTqsIIMzq+GT1/9gz/SYVJtCytVMXoeHTCTqvibC5Flp6GxuPRwShRhRWFxfa8LDOZcGx/9NngQzcjk8xFzvGTrUsBZNTks0gsPFhqVQptpWe5kCrjnlOX8ZN37hPPnrIzPhW2Zm1jlgoL1ETKJjWzc8GM698Bk1PGixq43JHrZSYWItvkfQaXtQSW7j5TU8dl1dmWgbFaFjP4LooqHx1/HMliquDrFdfYSmit9OgCmkubYfvfGxJvHOPlwagyQl/mwhg+E5G+WsJblfLWP8OsYjAYC1HjJ+wmUmR8oirb6TVwy4XlKbe9I8PTVHri79e0pQrxbgqytU8UVYZdb79PlMoEL+JLoI2J0PGVIFB0hWvN6/hC6MnyYtkV+6TtGmskYLJIsioFmsyUIpFTmUQGkiggwKZKg6lMqtkMm/EiZVZkAiho4LvcJvGUZ0YcgGv8WUktUhzVrsFEiowG15vp7yt972LRPLjhYRvj3txgU61ypquicBIx4z3D2/8YOCLPH/v8+BxMAUUSxrIrbkB9C57JlIGKcp6BpnIQ9Zv5tbHcViZqJ4HsDCyiK9h/zs/fM9/aDRoaaVDLLl94hc7FXOMDVMVcactMledc143LOgdkV3OZzJn0QmvJC+tMgxfhvuPjjM5ymd2Vs2+NrDNVgGxsXuW8ONQCDGOSCRY9s9Yd5CQLkLlI7ufxJ+zD6ZsgTAKdnzhfT9HvNdi0BNNkudKCIQjp3HvCCghaN1FJUmktgH2nX/dxiTDCzWi1yov8walpi/L0mkFjH07evx6zU1XN5lkNw5Nl6sqw03+9ZD//49nPY1SauOZ56S2jsS8w9RW/Fzxlp97EV/25o+UHz3n+S7/nJJkEr2BzeNzgB6sUPFzUDWGkCoYwkwsB/+aqgjaoW5kLcipyqkfmVM9/WfE7B0t/i5YePMvErvVfVYFPgIbEdSJEimpDZ+NWsEzm0o7Ze8gMoEMQcRmfcXASiI1CN5Hqp22RClIYyMAYqBNdSYB60LCqQlyXPvETQQ3kWuRaj8i1forj1Wu0bMxpz7y1r/gH9zrZNPlGGemYvavAEgTPXVSaCAg8hbKgMNEmgqmY8iqzPZO6WwiTU5FTPS6netbH7Q4u5LwETY3PDErj52/Y+c7kqyqcafk5Flg51xzmf0LjUxdeIxAZX6i0xm59K+9x/C5luts8MGLKFHii9+tmPD0BZxhVMfazf1cyiF3ST8LBFq3I+2f7W7qnVb97gvSQfh+7usYogqDVhbjnbytIysGEP7nbn0bY2yAJVnbAWwHbuC9pNEWS5RpA9eTq/majhvDYQeQzvnQAWYm4tq4soKAhMkglhy5uqCCqXdxVPf7udvXIHEWJ9bO6Egv0q8RvuSy4DRjVyNWIiSM3ZifZFa8N6NGR+TRy+JKv15S2dLXSPRUwDzOBlSOE8yVXeMEbkMjhDY8w2AzvHLt2RzfIbvnPCeR9f//x4AbvLMfsNwVJoZ1Da3we0jowTZc6ZqKY2bnPF814Z+2u0vJuDoFs/G1WKog8+CFUyAB3VKUBeO8Fdsg5Ajag78LFejUtHmgcnleQwWKRCstkTiEbffWXXaqLCbTXtVM0TjaD5e3q5zrJ146G85wO9aAQH/ecBPfRgJOxsatIFY7fSy/wazDHDRlPglixzfZU/u7rROi/az7UXniQCyEV70GO3m5rQxhpIOlrFOrYRPpEgrer00YsgzYxbMWSYr+DOuvfIF6hejy7HMKSM8mbu0IEoRmhGaHZDjTrogWB2sNBbbnsTJczAV6WxnQ23ih9RTJL+Eb4RvhG2Rpla4RmhGZ/CTSjbO37ZGvLi9vatMW9657iXlQFP8fS4NTVN+GDK4DL8DYlq3FV2JuM55ydWZ58Qc8ruTHs9OTfLBAfxxRyXmMNUotcLRyhphTdQk1lK+2WvVFpkOCZ4JmSzceFy5RsEpoRmlGyOShQo9Ig4Rvh237gGwEbZWuEZoRmw0Azytb2szR4n7675YmrhYKxgxihw28ne7oeVguuVqS6pYfuxz0JooT7qaU2bknn+A6Rb9cySVObRoeNmUSX+pdK+gZbF0v621+7XLIhz0COvARLG7PXU5ZD12SZiQ5xYKcFtlxAP9IDJFC7S0A04UAW/k+x1gp6LFIY0BrdPK7Nov+7AiwMoWIwvGEtZpfPE1yrHdb+HriF2x7kA36qjIECZeGMwzylUi2FSwqXlPw/vjhJyT+hGaEZJf+DAjUq1RK+Eb7tB74RsFG2RmhGaDYMNKNsbcil2jNXBNxerF0vEoY6JFon7qygq8QVA9FUw6rPxntzsOTsW1RvsfC4bjfNhf7KLd7eWrd1VcxmSwRXtuVZdqff/38o5J/V2hYAWCBFSR1NpOP5tBsuSLc53s7acNu+sZC2LB4EpSIrBToKdJS2P64IR2k7oRmhGaXtgwI1KrISvhG+7Qe+EbBRtkZoRmg2DDSjbG04RdZQUuwtt54rlTUHTexeIatFCQ4qgsd6d4VxxdEKtUxZLFTyzVbKcmMk7pe7bkPrV/urrm2bLaXXtiT8N7Nef73HstmN56k2SvGJ4hNl248tMFG2TWhGaEbZ9qBAjWqjhG+Eb/uBbwRslK0RmhGaDQPNKFvbz9roAYioyksteDh2qSt4c2CzKNJLNb20lT+w2X/L2+qnqqy7rb6IwvQcDeXos6t5vaoCMuRcgqXORCE0R68dM791qfu9/DE7ZH+cQRv/8DjqwR/H7rfyngz+nN3M1+g4Jli91Ku1meN+ckGEnRQnlW2+4iGLrOQwyHy1EJUdHrLKVLiv6kFT2MUqcvhVvzuYidfNGtOw4hTPVMIxlUUlGvtst01gVxLAsmkb1Uy7kqzpfk0QDa6PJ5qh9eNdkCKdCesOuUKil9idyPE2U/rb3XXrytuGfuFia8/MYMO+Jhrs8tIIrpM5XL5S2VTz/JJn5RzPmirnyirQv8OCVFy6XWxLjaeLxX70opIZaBMr9H2HauOJ2VzPqhxso3/GEtWeu6Lf+dzvTRrNgX63nvzdezz4+gniu2dc35r3xXKvVAHdcf9Fw3j5uRe57oXKq9c9K6uNLaXz8ucld8PSwbg30gRPC65vxuwVT+bR4nJ8j8IaCsyrc/ygXwGs43fUz5cKtwBBkrtfUSExLosGOJt3NE/aF0JPoxXx27Z4uUOKHfoepdjusLudMENn6NEEhCYgNAGhcsqezDyonEJoRmhG5ZRBgRq9/CJ8I3zbD3wjYKNsjdCM0GwYaEbZ2r6+/Er8hiebuEBn6BE8Ezz/NeCZcJmSTUIzQrNhoBklm1QaJHwjfBsqvhGwUbZGaEZoNgw0o2xtP0uDD1iSGTvINz1Drznx7sHrJekwPTpMj+Imxc1HHDcpYNIsgNCM0GwYaEazAKrZEr4Rvg0V3wjYKFsjNCM0GwaaUbY25Jrt//0wPX/43YOLuHSmHtVaKd5RvKPs/VEFOsreCc0IzSh7HxSoUa2V8I3wbT/wjYCNsjVCM0KzYaAZZWvDqbV+vzP1ouPvHlx3pfP1KFZRrKJYRZk3Zd6EZoRmQ0UzyrypTkr4Rvg2VHwjYKNsjdCM0GwYaEbZ2n7WSel8PTpfLxp2Ol9vVXvuik7n631/VdD5ent7vl4cjd6utlnZzNba7UsgRdOVzy7N3O0+Hd6GAbbyfvi5s+F1PafXorpmt2mt3uE2XcCZ4KZoruOISaZJOLxg9xxj/5A3yEZbo1Yt0SCfVsWZa3sKTwtj/SisSFldiSXSsHMFX0elMg7EuZ3Dt6PFD0eGT4Wtj2AcDj1XtCihcWCdOt2MYTS3tjw+OuJFfTgHEiI9zHC7mkOD29WME+VnWEYkkDjb2j34a5OifbzAe+ikoZMvVFq/8pvW4NCthI2iQis1kvJvPJtDE9u3jZ0E/Nquj2B4qt04aOO5qFH4xUYbzjZy+96n4neqPb1aS6w6+VMcfD/e9DGOwmc3NkYRbNW+i4dRzLuIPArtQxZT5ZwrWFG89dBZKRI5lf7VNJBFQ/D2vfihZ+4FgCb9i2ETP9hY/9rO84gZYVck86lgnT+cGeHgExTlJnHGZUdAReJPlbQHoPBDJvCwKU8QQpAo5IqfEMAyBUYPqLJJfAJmB7NPEAbhzvfqLaKc8XgDNp1zF0SCxsGxsEfBeTty30Sj3W0YYALN8ajMAFSRfph2e7f76DXpHQ8+RK4HA4XuhW1ubnA690FnyyVeBsfR6FfwccG15BMcuI8IQnPBUzyjFKzoi6gd8Lu+HZ77mcaCZ5Uz1rLMwugcOQtBQ/ZPnCSJKO2tbS8iEHn/7uwcGk/AjVEqjBZwVfMrFAb+Ph7hfMDn8W7BAl67GWW8mFWuSjHyNPHP/wA0niqR +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Run a shield. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/run-a-tool-with-the-given-arguments.api.mdx b/versioned_docs/version-v0.2.23/api/run-a-tool-with-the-given-arguments.api.mdx new file mode 100644 index 0000000..5ec2e74 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/run-a-tool-with-the-given-arguments.api.mdx @@ -0,0 +1,68 @@ +--- +id: run-a-tool-with-the-given-arguments +title: "Run a tool with the given arguments." +description: "Run a tool with the given arguments." +sidebar_label: "Run a tool with the given arguments." +hide_title: true +hide_table_of_contents: true +api: eJztWttuGzcQ/RVin1rAlpwgaRr3SbkUNZDEhu08FLYRULsjLWMuuSG5tgVB/96Z4a60utiS6wKti/VDLPEy1zOH5DjTxIEvrfHgk8Np8vLggH5l4FOnyqCsSQ6TgTi3Vh+ZG5tKGjoFX+nQS/aS1JoAJtAWWZZaxfn+d0/7polPcygkfQqTElCSHX6HNODG0tkSXFBRa0uMNXA8Sg4v5lt8cMqMk9nehrl7xMXZVQlsrUcliSrkGPB7BiOJjiyNtP3+oOhboYwM1gmSJuxIhBxEbbBQAYqeGOhbOfHiMoq5TNDYWuJWzyunN8X76+mnRhULEqg/k0HyhDI8MbKukIGW0cwhr+tPSd3st6H08MurvSnNzHriiw2Ae3A17U9tpTORyxsQGsw45EKjk8FTQrdZqzbFddn6c7SN1MRpUVplAn0Ils1GuNnKpZDMMEoyyxRtk/qkpWgktYc9BOaPSjnIMOOs+QrNU0GTapRPQSb37kk0peejSW1GI4dJDMiarXFYAC2ErI41B1p60SBvm53LIo9YBm6XYkV6HRBMpTSCPXhMBNjHBletULC+99HhI4Tjmo+D2q02ZrmcnlZEAe7CUg3NB55UQiQlVhDL2442XNXI+lsBZTWteJLAh8MZWirraOL+rO0p2V2Hc/JFFiS3VlcgTzIkp7tSRMdmHZt1bLaFzVbZoiO1p5MaKZ7nBSecBqyyrC2ndYxI5+SEUorjfuNNruO3jt86futua/86sXW3tY7NOjbrbmv/K1Lb5bY2u3pozZqen47L6MbPggrLVqGsFmpHzhYcxmCtFnAHacX70A5wzrpvBXi/TET3BLCl5yPtFPVOoUYbFIiRVBrjN9dDCG8pwZKHMbiHtHypCnAqFbxf0P6tqgoIcqXk56jdnO8NN2BTad0+zIeoDaRpD5mqGKLxrZFWA3Tpqr1+KZhdzR5wejC3UjTOCDnEpG5M4jYYz7G6oTO8Fvo4TGUnoyI139Ej5KKyV5saz+fM2T8q8EHcIq3hNqlVRjRWSE2HECbnH+tC+yBD5XdA0h/n5ycirmbwJIvS2wb1iG9evIfB8Ll1QfiqKKSbNLQUUXmbqzQXKjrtlKSKi/Qd51kP5zsgSHfVHFeTam0N+ibyqpBm3wHiYajx/FnsWTKHrxXIutKkjyvoeJ+IzqRo/BBE5fFAwhPZAe4GvAkU1mEBmnipIM0LWPoSUjXCUrVpWjmHxxms2PUYvq0z3GRrHrwWLXKY7gle85eSBfUNTo564tRW41xPMD1a21svTn9/L978evCmxxR1J4syIqPBF0J9oe+dzMRphPjCnh2RX1fOy7ebKyfVing6p+sEfcDKw81m0ggmqQiEsboB/F3YynCJBlVAV1RdUT2zonr5dulQEp8J6XVl+XZp/Wmr+ACBuxSALsf8NJD4TOG3SE+c4KGMBgVEjxxLLBKNk645qV7fd1J5cDeYfLpyV3S5QsmYm8rAHcY74Deow9CVVldaz6i0XrfPK3420B3yLKJ9ob8ur8E65JtgZD1xXCESQBZ8Kg0BDx5jAwYM5+rymr/m1joTDwjuiqorqudVVAebtO1QQlwlBDU59uQNvb9O8bihS9vV2rurMs2L61aFnK2tr3tujG9gExtvpXQSn4TgSOJVjBaemu9sNln53zlPqSwy45vh9uoubTxa2WSZPUCw0LPxGjik17fogt+k9T/8Gh+ITKVsWyz0eRbIuVJ63zQqyeHeI3tJ8/DOg7PU7aHIMVjqxwajaCEguApimyO3+DUpLT9IShly/Na/edEnBfsuQq0fM4EL4qXHc1y5oZzkIZSH/T5ev/ZzFALZvtaykPuI/PS6l9rYgPeAeFZhwhs/NJx/cUVzhKzTBQY/LqqnhaEFduZQmDE/jSyvrB3/RLrFGekWZzWPMH5xIxkeE3PzYgMGiXb9EgHJNiW2JXN/G+F+o1DIpRErP8jxwG8rMBn3pD1SON87FbWfXOQUahcLOm7cSKbgo1Dp4NIQL1jHhLkufEhvRI0U6qhlFq36jAeS9lwolEhkd4pJHbgdWWEpHC0W2HV/XQ3UquyXGq/RZE39N4cIq4sY9zawqMscoYU4IPjQqumUutlfnZ7NaBiB4SaRqfhQHFKiL6jacjzBsFwJUteAS5K6s7l/Hhut+IauuGRXaYyqN+4YpCmU4cG1V60yOTk+O8fFw5oqC25FJk7e4iD9e5jgB8tBZBLisWmipRlX3BtNokz6+Qv9IIV4 +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Run a tool with the given arguments. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/run-an-evaluation-on-a-benchmark.api.mdx b/versioned_docs/version-v0.2.23/api/run-an-evaluation-on-a-benchmark.api.mdx new file mode 100644 index 0000000..2712595 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/run-an-evaluation-on-a-benchmark.api.mdx @@ -0,0 +1,68 @@ +--- +id: run-an-evaluation-on-a-benchmark +title: "Run an evaluation on a benchmark." +description: "Run an evaluation on a benchmark." +sidebar_label: "Run an evaluation on a benchmark." +hide_title: true +hide_table_of_contents: true +api: eJztXetv3DYS/1eI/dIUWK+dor275h6A4yRX9/Iw7BSHQ2JsuRJ3l7FeFSXbe8b+7zczJCXqsU+7hZ1jP7RrSSSH8/iRMxxO7wa5UFmaKKEGL+4G3x0d4X9CoYJcZoVMk8GLwce5YF/SCSvmvGA3XLEgF7wQIStSlpcJPBdMXPOo5NhgNBgOgjQpRFJgVzzLIhnQm8MvCvu7G6hgLmKOv4pFJmCEdPJFBAU0zPI0E3khNTUw6FiGzneqyGUyg++aBP6SyN9KwWQIY8qpFDmbpjmRBT0MlkNox4tS9XUkkjIevPgEFMdZJGBS8EwmY6BjBoxR8NeUy4geI9VhqX8HPAlEhL8v28SclHkOdDBxK4ISnzE9OkunNUlAEw9Dia95dOZMesojJYYglN9KmUP3QJrhQjULGLGQRYTT+Bn6ao9/TLKqR5cJtANq2Y0s5paWIufBFXBgBKQgMd+vkjsSIpQWu0xAyDJkwNuYR8DiWPPiYYTdkZGEbmci70zwp48fz+w8gjQUKGDDkE2K8jrPUTHw4yHjTM3TvGCqjGOeL6x8BH1zM5fBnEk96VxyECjqFE/Mexpnif0XoB/bjqy/xqGjNIG5sXkZ8+QAzCnkk0gwp02DHBzJynHzWM8+ZFqxvmXH7Jfzt2YyoLNsIlipjOUKaC2uBYvTHGwnQXmS5BifpGVBg6tMBGBQAUuDgNQa1KhJ1y6KbCRspVUxz1FoYtMK5lmgYtM8jYmG47PTETtPy9k8WoB4oii9Uez8zQn781+O/jxCpolbjnbt6heoej3eSx6yc63iNT1bar6xnO9+7LecIJIIA3NorvBHkabQOFnYjrFXUISZvBbw3zgt4RvkrYyFNypvVE/MqL77sR7vI2j6O9R0Y1nKNa3/pCXYBHBI3AZChMg2NDbYT7BIxrIYsbNIcCCoAO3hMw5GEsHL3K5UP6xaqZTIr0H4wE40JQHsQcUqE3EL/MbdijBs8KblTesJmdYP7np1ipoNlLMLre31+Ma8jrsqb5kRjtiHEjRB8JhWpYmAhSdJacccWvMKxZSXUdE1sXUde6PyRvW0jOqob7QtTIisBFWNzxTO5jXsC7su4Dk4xagklU/MkKHA9ySYg5ZdoYuc8ZzH4HDm2M/dIIE/oGX1ifb4JPaW8WLe4QaueKevrByqZv0uOdNeeS2GIi/FsGuURpWWy0v9MazdL9NwgV88kH3X84MOp3K2YutM78pcE299+Qb71g+DUx+DqocyhL3DikHsa+SZ4ZXArtNEfJiSUNYPot927RCoB2/ixSAGWIoGDqSaJ6A++sdGG0Yy6VMUdYNKjGmgUkOrMWmS2gZbcZc1W6zYPpnumP3sIVkxy2Grt2jwwnnkkvKuBJcLAOqz+eDzACduIjsL0ANA4w6puyEQEenAzT9poAvT6YXts02Z/qw7uA6LKREBYxQp6lzO5ug5Ao8mfCIjWcBH6RW6eQUTPMAojMhQiPfla5Fm46zB1vpJP1fp/ZZMBRaJGKgByMxdMpIynvQsyScAEXkaKdjMJ2EaJ0KRi2s7HrGfkC85QyUW+ArDiEo4n9OQNIHewcwcj0Y//tANusUluAngSze5Pof1Yp5GIYFIUgaRKJVD0Svdp0J2YLf3U6SPaXa2UY3go4OMPTO0fLtSn2jJUvUyp2IeRahVSug4ASqU0kG9oH/2//g7ezAtu+po2dUGLbvaXsuot817vvekCnrymWUADIF0wjA66lsL15ID3ESPsmDP9xCvpa4p5n9tJearVcIFpQQeI1xUH8A0UMpXgPdAaSSvRGRAQ4EIcYchsftYJryA7Qowy8hw8V7vHAy1MazOKD34wCDsxnXBg/d+4N2BKg/iHsR/NxDvoKQH8ycK5ss6VDHoDI3OAb8d649XsNGaUK+7wG9lXMYscdirWYu0Gu9/JhIEBPBupXYVzeknnt4y7EUDXmDPJBZpmaNSxhngYgTGV5OIXSZpYaK51Bl5LN9gWAZovkV5JbNiTjqciwxGwXHGmQChFYu1IPF8hdJMRHGDwauD70ZH4GiHDP47YmepkmQ+BptoBPlfwRJxY7kw4Rj4AKfyZi4KxDH414LBmil4bplBRKuUTXk+tABHUnXmRjKV8zSlKErBoysTMNFjZTJQ2kmD384UeZ5zVC8JmNxzHL3sHGyjZbDvAS7AGU8CmBTQnQsb9wDgiCKGY1iRIp3TMtdT0wrH9GEWwH+CAsLJUTOUGoqIm2lTN3ag0a5BHaO+l13VPtPuaXtq9jWrAyGaYwtYX+NxDCsJn4nNeAeL1Vq80/01AM951I94+oM+yOMKI4b0mlkSl43oZ280A1/aWI3t3hjU58GInU4ZYH8hwQZbncNwOa0H10BHODTqmlOnAbA8QSvWIi4TQM1ogSx9G/GYs4uCB1cUHAV7hVYR6DRIlKWkHe1xniHemkjZkMBXxwILDWhpxICBMiF9UN/2BgcqLb57sMCBjJHDrvDqJy6fX7m7Y4rLWnZb7qPJjdhxdMMXCmRA3XweUEw13krRyjzqiYpTaNUMRR3hGXHIC04vjGlpTuJn+OYFfXd4h8Mt/4qI9Kfvh3f4Zjli71MMSCFWY3tAYNj20HGZRlF9RqY2R8DKXG4XX8Jh9GuWpbDKOOsXrGwA/IHYDQtwZAcGoH/CNZjeCkGjeF4noKfktww0Qzq06sd0vIeHhprXxGg8YLeByw10Nrs8pT7IpFu9G4boWD/NYI+9htYrhxU03ome8GkfAh2babk6+xD+O2B+c8dnH9zLhLAXbUHU32Ztw6XHIuVemzccxt27wd/r2Vk4Qxpu7uVKbwkRHs08mnk024BmbbTwoHZ/UHNdSkoRAB/5WoRuP84y0vFC/G7N45vHN79be3zA5ndrHs08mvnd2lcFatvs1paX677ZjXYKkdZBSjdGS0HAdyaQ2SW+GSQ0cUjUTsxDzMuAgoGoqjbWb6yXwuQ7xpAtjpusqXZ6k0P1O/zkpMru6pKtc6bqBC8KbToX1O6/SAM3kqZC6yc6FrxjYhs1rcndDLA+9evpZw/4rAGfNeBTv3y2gE/9eoLg7VO/PIj71C8P5j71y6d++dSvh0r9kklWFmM1lyIK1Q6sg6bA/73bYnLTLE/LbG27ddlO62FZX+3rETjPZ72+ez/ve0hIyihyCZjATARP3EfGspwnXbr1XLsTWV4ud1IFmqiZlqMNxxje+Aik/ROZ/G9Y8I7xi2XfJyQQXTljjHLZRpQ7sb0DK/CVDTIWREDri00dOPdsf1p7j/cGUdgOw8JU0HbJvQb6x820GrVBwnjVBqK11jhx2aoteyZGs9HQRK8RPGnR+nZnfm7JxHpgy8laEzv24Kwr+uZrc8h/VwsC7GzqjuEP26cO1aHcZHKd6ruvembVbfk/3jjXaKLZF+vlsIozOlOb0hpg80p3Q3xj5i21aVLjtGjsuNLorFK9Nv1ntUirTFN9355YP+qukc6M30pF+5XanrQ4SWg8CERWkJbAO946D3qMuLtmpscVlcxOxrmsX1gQ3U2cLSGB/vTshBs5wLqEAGwgctgpyljYPZ8aVYvqOJinsr9WgS27xssidZVlOEiA221yTnQ/K+0WKQNyGhaLegOdYzWDADdCWMUBT9vmMmkeU5jCCJgijTYxEaaIAOZVnxYwY9hMhgo3jtjktD74YG/oO9zIBDzTThsw2UKjOQNBorNcYLK2uXhvuaN31WN9QLuOSXTJfjiYlgmNOy44vs0WxRz+iEDxOzZGPb/RHXfMTG/mzbkw2lcAninNAhwt8FYP2X9FnmJdDCpaFal1k6hPOzb4pk19WLWb+n/SjB5suGycKGkOrASCvklWExtW8x3ihJF3I3YCPiBdA9AlPRYGWRlCAM4cO+F1vQ9C3UaMpWb5CEcywYkH1+WVU7Y8Vo6Q5im6XVafaT6o0OD9vFwww81h40oE+UC8wOBYUU/bdoA7DInVcWBBAXTNTKHLWnDsgP3atrIRzuvXF8aDtiTQTQ1zjQLLUaGW/Xzx4T3TRtLfk8ua/h7BS4tSg76c/c02+HvVEuX5D3BNZ/0jONzuH0B7U0jvGX3K1AKcxVt2cIBFa8xKa0cz3O7e3RlPxJxfSx2yXmnmGRoRGXoW8UCsFf8JgQ1hlpF6ei3yHLYwJB4jbNa4X0MsaJxzvzRkjfTYwILjTFsybZLMnqh99m1UYNcxzLRgkHP9S63uRgcD1xChHfoWYRjYkaARpeGCyR/55u6u0gfn5s5y+Q1F85IQS7gIJ6hQydP5GhcBzAKaYG0gBWiOhaIaGKXFt3HD0cBsvWT0hJ/bZ+MG2EY2NCaTKWw3Zctb6omPPT9a7lFkxakfa1ChcUJv6y/ZvIftujbSaiRMdLrdzoOzhqDxutHJsNqOFSISsC/E4nlJWE2pdllEgn7VWIH26FhcrsCgRbOmVI/j1JvFVFE0jfjMKhVqn42xmWF0shQWHQPGgjLZUUPtuOmSTs7qsWKUuviTxupGOsUD5Cy4i9TYlDQaDmbgUcQ87yJTHSd3GvRFy+3Sov0kG45vjlLblPsc2OP++bScFlR/Wu409TavT4vQIAuKELgDPIM9FS5nZsW5ePWvoWaexH1TgYdw7NdsEXLga/DrfTKMXIY628mf4fEFPT03NK7YQHeBypnkwawk4K7maaOz98s72lszzSfba6Xts3nmq58BsyfJ9Alq4cv3b5iZRLW91eJbr5N76ReyqHFMTePurFSG3vUKtVd2wAZE8VjpsfIxYGWlTH+8jnrU/KpR0z2Gb42wE+k2V7nhFbSPtYy7sw9PTHCt0+Pq1OeknUzczX3ea8lo+1K+JKZPrfN50T6l7qmm1PlUuqeeSufzoh8hePu8aA/iPi/ag7nPi/Z50T4v2pfE9CUxfUlMX5bElyWxew1fZOkx1iPZy5X2RZY8mnk080WWHiuo+ZKYHt88vj0OfPPA5ndrHs08mn0daOZ3a74k5ldYEtNUsLy3XvvSmD6LwKeA+eyBx5494LMGnnrWgE8Be4Tg7VPAPIj7FDAP5j4FzKeA+RQwXxpzLSz70pi+NKYvjbktP31pTF8a05fGXDNTXxqzY7ePuQAiEu1LY/rSmL40pi+N6Utjds3cl8b0pTFXlLnxpTFXlIj0pTF9acwn4LT4cm++NObj0MKvtsjbXtkBvjTm49RSj5W+NOZj0sevFjX/r0tjupN/DS/q3tCDD9J8fWLx1jq7wVat5URRPOZq/KUMW5duOi/a6lmYcyhDc+3A1mcAQ8b1lYFWZ6uAovVZDn7lLfJCkXs54UoGjXNuPfKbRB9048EYrULYfLyl+3laO50mPPj27TvjkNrYFMaNqFOdGaNHpaM7HXbDSFaEEtzltPJER3tN6MF2UbmVeryqcqmeEw4txsQWsc2B5arkhnPdA9U2vS1yHujDC56oG0z/wZypOlHF2rBO35/B6KTKYyvvHQhx4k7X0DsJ+UbI2RzGGdePYhFK8n4xRDIDbgccI90lpfzzIADACRZNs6zIemOo+tg9FNNHpxQGduZR6a1ypYuzxqhrz6nXcW9bYCXs8CjZSIeBUFQ0FmUu5unNfhsAR5f7tWCVUBz2gEYfq5+xbctiVp8BamZAwwOuDrQuduy84f7v4l1Z7GnZd409nRf3wZ5WZ78v9mADBPAHslFvoY/eQtsS38IeSbRnpFo7GiQNwrRW/g4GqbXctcT6yX1M0Pby+9qeV/0/WPU3KvpLlNyOKk7S3qzbe0VeGnq2tVvrt8p+q+wX4kePRk9qq9w02J2hyO+c/c75yRvsY945G7Xd2TD9DtqbwBPZQTcuw7VG6HDlHc+qu1+dnmVIyS5Zkwx9Y7f97SIt2Q3XmZ55SSkxoDNjUwF5VV7Rio0hWlF93c720fhfNmGRZzfXHi0qqj81N7AwPwYvhlJq4kRUzXdMzcdm48A5HmjF813BiiSYxzy/2ud8YmIbj7tHFOdlgscK53idTBVazHVbkzsMYpqn8OcgSxWBGi/m8Nfh9XMeZXN+iBM5rEZRh3f1iDJcHn5JJ6iFAC7XlAv2yZTKGsyLIntxeMiTxcE8xQSjgwgTPw8UJn6OglSXFlMCbFziDURo+Moi5adLfIdnhYb2l2m4eK0FhWrRmTQ8azG8Vh67DTSuTY1QPbVb3Oosprm5Zb9sXgk96r9G+dxeNvxkh7nsJGNi37pkel0Gvape7t6G65wAdQzkaEmJcNOU6LX7Vie/9sI9K4RxUEbadK6f9yxE+tyaQK1xyGh9PKdnSs41lqQ+J6z1D7eXq0USUmExRbgAvUgsVZTr3ZYpZAQ4MTX5mNApz8VnvBAZpTnl+3Y7p4TgSOhVxFBFFYk0iKMixzyp72OhIWApr/rwDTPFOavUaNTmxZ0jkW0aG13BWxeH4JFKAjNTMk4b1KeBMShcGqErXIoro4I/XrhmhU4L2hUYAdoOtr67w1uzv+TRcomPwSpyNBr4ec1ziYmEZEKhVPg77M8SdKf17NwggcbO01dWyhUhBpfpYWP2K2ds9wYJ7gzoxhP8BT+vxAK3P+4M6aB7LniIPg8Qrj8xxbUOzJpuu8D11ujiIaWtYzxTtzimC0Vrv710QO7sw8VHZH2qa3IgJqCrxG9wpwT/JnLTrNrt0LO7QcSTWUm2q9Pml/jP/wCfo1ng +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Run an evaluation on a benchmark. + + + + + + + + += p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"system_message":{"type":"object","properties":{"role":{"type":"string","const":"system","default":"system","description":"Must be \"system\" to identify this as a system message"},"content":{"description":"The content of the \"system prompt\". If multiple system messages are provided, they are concatenated. The underlying Llama Stack code may also add other system messages (for example, for formatting tool definitions).","oneOf":[{"type":"string"},{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"},{"type":"array","items":{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"}}],"title":"InterleavedContent"}},"additionalProperties":false,"required":["role","content"],"title":"SystemMessage","description":"A system message providing instructions or context to the model."}},"additionalProperties":false,"required":["type","model","sampling_params"],"title":"ModelCandidate","description":"A model candidate for evaluation."},{"type":"object","properties":{"type":{"type":"string","const":"agent","default":"agent"},"config":{"description":"The configuration for the agent candidate.","type":"object","properties":{"sampling_params":{"type":"object","properties":{"strategy":{"description":"The sampling strategy.","oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"input_shields":{"type":"array","items":{"type":"string"}},"output_shields":{"type":"array","items":{"type":"string"}},"toolgroups":{"type":"array","items":{"oneOf":[{"type":"string"},{"type":"object","properties":{"name":{"type":"string"},"args":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}}},"additionalProperties":false,"required":["name","args"],"title":"AgentToolGroupWithArgs"}],"title":"AgentTool"}},"client_tools":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the tool"},"description":{"type":"string","description":"(Optional) Human-readable description of what the tool does"},"parameters":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the parameter"},"parameter_type":{"type":"string","description":"Type of the parameter (e.g., string, integer)"},"description":{"type":"string","description":"Human-readable description of what the parameter does"},"required":{"type":"boolean","default":true,"description":"Whether this parameter is required for tool invocation"},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}],"description":"(Optional) Default value for the parameter if not provided"}},"additionalProperties":false,"required":["name","parameter_type","description","required"],"title":"ToolParameter","description":"Parameter definition for a tool."},"description":"(Optional) List of parameters this tool accepts"},"metadata":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"(Optional) Additional metadata about the tool"}},"additionalProperties":false,"required":["name"],"title":"ToolDef","description":"Tool definition used in runtime contexts."}},"tool_choice":{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model.","deprecated":true},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"title":"ToolPromptFormat","description":"Prompt format for calling custom / zero shot tools.","deprecated":true},"tool_config":{"type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."},"max_infer_iters":{"type":"integer","default":10},"model":{"type":"string","description":"The model identifier to use for the agent"},"instructions":{"type":"string","description":"The system instructions for the agent"},"name":{"type":"string","description":"Optional name for the agent, used in telemetry and identification"},"enable_session_persistence":{"type":"boolean","default":false,"description":"Optional flag indicating whether session data has to be persisted"},"response_format":{"description":"Optional response format configuration","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"}},"additionalProperties":false,"required":["model","instructions"],"title":"AgentConfig"}},"additionalProperties":false,"required":["type","config"],"title":"AgentCandidate","description":"An agent candidate for evaluation."}],"discriminator":{"propertyName":"type","mapping":{"model":{"type":"object","properties":{"type":{"type":"string","const":"model","default":"model"},"model":{"type":"string","description":"The model ID to evaluate."},"sampling_params":{"type":"object","properties":{"strategy":{"description":"The sampling strategy.","oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"system_message":{"type":"object","properties":{"role":{"type":"string","const":"system","default":"system","description":"Must be \"system\" to identify this as a system message"},"content":{"description":"The content of the \"system prompt\". If multiple system messages are provided, they are concatenated. The underlying Llama Stack code may also add other system messages (for example, for formatting tool definitions).","oneOf":[{"type":"string"},{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"},{"type":"array","items":{"oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}],"discriminator":{"propertyName":"type","mapping":{"image":{"type":"object","properties":{"type":{"type":"string","const":"image","default":"image","description":"Discriminator type of the content item. Always \"image\""},"image":{"type":"object","properties":{"url":{"description":"A URL of the image or data URL in the format of data:image/{type};base64,{data}. Note that URL could have length limits.","type":"object","properties":{"uri":{"type":"string","description":"The URL string pointing to the resource"}},"additionalProperties":false,"required":["uri"],"title":"URL"},"data":{"type":"string","contentEncoding":"base64","description":"base64 encoded image data as string"}},"additionalProperties":false,"description":"Image as a base64 encoded string or an URL"}},"additionalProperties":false,"required":["type","image"],"title":"ImageContentItem","description":"A image content item"},"text":{"type":"object","properties":{"type":{"type":"string","const":"text","default":"text","description":"Discriminator type of the content item. Always \"text\""},"text":{"type":"string","description":"Text content"}},"additionalProperties":false,"required":["type","text"],"title":"TextContentItem","description":"A text content item"}}},"title":"InterleavedContentItem"}}],"title":"InterleavedContent"}},"additionalProperties":false,"required":["role","content"],"title":"SystemMessage","description":"A system message providing instructions or context to the model."}},"additionalProperties":false,"required":["type","model","sampling_params"],"title":"ModelCandidate","description":"A model candidate for evaluation."},"agent":{"type":"object","properties":{"type":{"type":"string","const":"agent","default":"agent"},"config":{"description":"The configuration for the agent candidate.","type":"object","properties":{"sampling_params":{"type":"object","properties":{"strategy":{"description":"The sampling strategy.","oneOf":[{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}],"discriminator":{"propertyName":"type","mapping":{"greedy":{"type":"object","properties":{"type":{"type":"string","const":"greedy","default":"greedy","description":"Must be \"greedy\" to identify this sampling strategy"}},"additionalProperties":false,"required":["type"],"title":"GreedySamplingStrategy","description":"Greedy sampling strategy that selects the highest probability token at each step."},"top_p":{"type":"object","properties":{"type":{"type":"string","const":"top_p","default":"top_p","description":"Must be \"top_p\" to identify this sampling strategy"},"temperature":{"type":"number","description":"Controls randomness in sampling. Higher values increase randomness"},"top_p":{"type":"number","default":0.95,"description":"Cumulative probability threshold for nucleus sampling. Defaults to 0.95"}},"additionalProperties":false,"required":["type"],"title":"TopPSamplingStrategy","description":"Top-p (nucleus) sampling strategy that samples from the smallest set of tokens with cumulative probability >= p."},"top_k":{"type":"object","properties":{"type":{"type":"string","const":"top_k","default":"top_k","description":"Must be \"top_k\" to identify this sampling strategy"},"top_k":{"type":"integer","description":"Number of top tokens to consider for sampling. Must be at least 1"}},"additionalProperties":false,"required":["type","top_k"],"title":"TopKSamplingStrategy","description":"Top-k sampling strategy that restricts sampling to the k most likely tokens."}}},"title":"SamplingStrategy"},"max_tokens":{"type":"integer","default":0,"description":"The maximum number of tokens that can be generated in the completion. The token count of your prompt plus max_tokens cannot exceed the model's context length."},"repetition_penalty":{"type":"number","default":1,"description":"Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics."},"stop":{"type":"array","items":{"type":"string"},"description":"Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence."}},"additionalProperties":false,"required":["strategy"],"title":"SamplingParams","description":"Sampling parameters."},"input_shields":{"type":"array","items":{"type":"string"}},"output_shields":{"type":"array","items":{"type":"string"}},"toolgroups":{"type":"array","items":{"oneOf":[{"type":"string"},{"type":"object","properties":{"name":{"type":"string"},"args":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]}}},"additionalProperties":false,"required":["name","args"],"title":"AgentToolGroupWithArgs"}],"title":"AgentTool"}},"client_tools":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the tool"},"description":{"type":"string","description":"(Optional) Human-readable description of what the tool does"},"parameters":{"type":"array","items":{"type":"object","properties":{"name":{"type":"string","description":"Name of the parameter"},"parameter_type":{"type":"string","description":"Type of the parameter (e.g., string, integer)"},"description":{"type":"string","description":"Human-readable description of what the parameter does"},"required":{"type":"boolean","default":true,"description":"Whether this parameter is required for tool invocation"},"default":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}],"description":"(Optional) Default value for the parameter if not provided"}},"additionalProperties":false,"required":["name","parameter_type","description","required"],"title":"ToolParameter","description":"Parameter definition for a tool."},"description":"(Optional) List of parameters this tool accepts"},"metadata":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"(Optional) Additional metadata about the tool"}},"additionalProperties":false,"required":["name"],"title":"ToolDef","description":"Tool definition used in runtime contexts."}},"tool_choice":{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model.","deprecated":true},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"title":"ToolPromptFormat","description":"Prompt format for calling custom / zero shot tools.","deprecated":true},"tool_config":{"type":"object","properties":{"tool_choice":{"oneOf":[{"type":"string","enum":["auto","required","none"],"title":"ToolChoice","description":"Whether tool use is required or automatic. This is a hint to the model which may not be followed. It depends on the Instruction Following capabilities of the model."},{"type":"string"}],"default":"auto","description":"(Optional) Whether tool use is automatic, required, or none. Can also specify a tool name to use a specific tool. Defaults to ToolChoice.auto."},"tool_prompt_format":{"type":"string","enum":["json","function_tag","python_list"],"description":"(Optional) Instructs the model how to format tool calls. By default, Llama Stack will attempt to use a format that is best adapted to the model. - `ToolPromptFormat.json`: The tool calls are formatted as a JSON object. - `ToolPromptFormat.function_tag`: The tool calls are enclosed in a tag. - `ToolPromptFormat.python_list`: The tool calls are output as Python syntax -- a list of function calls."},"system_message_behavior":{"type":"string","enum":["append","replace"],"description":"(Optional) Config for how to override the default system prompt. - `SystemMessageBehavior.append`: Appends the provided system message to the default system prompt. - `SystemMessageBehavior.replace`: Replaces the default system prompt with the provided system message. The system message can include the string '{{function_definitions}}' to indicate where the function definitions should be inserted.","default":"append"}},"additionalProperties":false,"title":"ToolConfig","description":"Configuration for tool use."},"max_infer_iters":{"type":"integer","default":10},"model":{"type":"string","description":"The model identifier to use for the agent"},"instructions":{"type":"string","description":"The system instructions for the agent"},"name":{"type":"string","description":"Optional name for the agent, used in telemetry and identification"},"enable_session_persistence":{"type":"boolean","default":false,"description":"Optional flag indicating whether session data has to be persisted"},"response_format":{"description":"Optional response format configuration","oneOf":[{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}],"discriminator":{"propertyName":"type","mapping":{"json_schema":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"json_schema\" to identify this format type","const":"json_schema","default":"json_schema"},"json_schema":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The JSON schema the response should conform to. In a Python SDK, this is often a `pydantic` model."}},"additionalProperties":false,"required":["type","json_schema"],"title":"JsonSchemaResponseFormat","description":"Configuration for JSON schema-guided response generation."},"grammar":{"type":"object","properties":{"type":{"type":"string","enum":["json_schema","grammar"],"description":"Must be \"grammar\" to identify this format type","const":"grammar","default":"grammar"},"bnf":{"type":"object","additionalProperties":{"oneOf":[{"type":"null"},{"type":"boolean"},{"type":"number"},{"type":"string"},{"type":"array"},{"type":"object"}]},"description":"The BNF grammar specification the response should conform to"}},"additionalProperties":false,"required":["type","bnf"],"title":"GrammarResponseFormat","description":"Configuration for grammar-guided response generation."}}},"title":"ResponseFormat"}},"additionalProperties":false,"required":["model","instructions"],"title":"AgentConfig"}},"additionalProperties":false,"required":["type","config"],"title":"AgentCandidate","description":"An agent candidate for evaluation."}}},"title":"EvalCandidate"},"scoring_params":{"type":"object","additionalProperties":{"oneOf":[{"type":"object","properties":{"type":{"const":"llm_as_judge","default":"llm_as_judge","description":"The type of scoring function parameters, always llm_as_judge","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"judge_model":{"type":"string","description":"Identifier of the LLM model to use as a judge for scoring"},"prompt_template":{"type":"string","description":"(Optional) Custom prompt template for the judge model"},"judge_score_regexes":{"type":"array","items":{"type":"string"},"description":"Regexes to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","judge_model","judge_score_regexes","aggregation_functions"],"title":"LLMAsJudgeScoringFnParams","description":"Parameters for LLM-as-judge scoring function configuration."},{"type":"object","properties":{"type":{"const":"regex_parser","default":"regex_parser","description":"The type of scoring function parameters, always regex_parser","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"parsing_regexes":{"type":"array","items":{"type":"string"},"description":"Regex to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","parsing_regexes","aggregation_functions"],"title":"RegexParserScoringFnParams","description":"Parameters for regex parser scoring function configuration."},{"type":"object","properties":{"type":{"const":"basic","default":"basic","description":"The type of scoring function parameters, always basic","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","aggregation_functions"],"title":"BasicScoringFnParams","description":"Parameters for basic scoring function configuration."}],"discriminator":{"propertyName":"type","mapping":{"llm_as_judge":{"type":"object","properties":{"type":{"const":"llm_as_judge","default":"llm_as_judge","description":"The type of scoring function parameters, always llm_as_judge","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"judge_model":{"type":"string","description":"Identifier of the LLM model to use as a judge for scoring"},"prompt_template":{"type":"string","description":"(Optional) Custom prompt template for the judge model"},"judge_score_regexes":{"type":"array","items":{"type":"string"},"description":"Regexes to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","judge_model","judge_score_regexes","aggregation_functions"],"title":"LLMAsJudgeScoringFnParams","description":"Parameters for LLM-as-judge scoring function configuration."},"regex_parser":{"type":"object","properties":{"type":{"const":"regex_parser","default":"regex_parser","description":"The type of scoring function parameters, always regex_parser","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"parsing_regexes":{"type":"array","items":{"type":"string"},"description":"Regex to extract the answer from generated response"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","parsing_regexes","aggregation_functions"],"title":"RegexParserScoringFnParams","description":"Parameters for regex parser scoring function configuration."},"basic":{"type":"object","properties":{"type":{"const":"basic","default":"basic","description":"The type of scoring function parameters, always basic","type":"string","enum":["llm_as_judge","regex_parser","basic"],"title":"ScoringFnParamsType"},"aggregation_functions":{"type":"array","items":{"type":"string","enum":["average","weighted_average","median","categorical_count","accuracy"],"title":"AggregationFunctionType","description":"Types of aggregation functions for scoring results."},"description":"Aggregation functions to apply to the scores of each row"}},"additionalProperties":false,"required":["type","aggregation_functions"],"title":"BasicScoringFnParams","description":"Parameters for basic scoring function configuration."}}},"title":"ScoringFnParams"},"description":"Map between scoring function id and parameters for each scoring function you want to run"},"num_examples":{"type":"integer","description":"(Optional) The number of examples to evaluate. If not provided, all examples in the dataset will be evaluated"}},"additionalProperties":false,"required":["eval_candidate","scoring_params"],"title":"BenchmarkConfig"}},"additionalProperties":false,"required":["benchmark_config"],"title":"RunEvalRequest"}}},"required":true}} +> + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/run-preference-optimization-of-a-model.api.mdx b/versioned_docs/version-v0.2.23/api/run-preference-optimization-of-a-model.api.mdx new file mode 100644 index 0000000..703105e --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/run-preference-optimization-of-a-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: run-preference-optimization-of-a-model +title: "Run preference optimization of a model." +description: "Run preference optimization of a model." +sidebar_label: "Run preference optimization of a model." +hide_title: true +hide_table_of_contents: true +api: eJztWutv27YW/1cIfUqBPJxuu9vyLUvXezu0i5E4GIYuMGjpSGIjkbokFccL/L/vHJJ6WH4kLrIPBdIPcSKR5/07L/cx0mAqJQ2Y6Owxejsa0UcCJtaiskLJ6Cw6Z2Nl7ERzIYXMflOz4+gwipW0IC2d5lVViJjT6ZMvhq48RibOoeT0m11UgETU7AvEFi9WWlWgrfAMv6jZtK5F0jtprEY20XJ5GPEkEUSWF+PerZQXBg5R8P/XQgNe/dyRuT2MrLAF0RkIjQSJ5PebNJzkwIgcGMvm3DAh73khEqY0K3mRKl0imxfT2Vhua9M7J5BsBhoPror1v8lkzPxpFqsEomWr3dBaw6u/ao3Cu8OHjDOTK22ZqcuS6wVTKbOoMbgz81zEORNeaS24tAwVZlyG947PkuhbLorncvaniXWhJOrG8rrk8kgDT/isANa7syIOcRISdZbxM7Q8uKx8eLxh5+zm6mNQJkbhZ8BqAwmzCj2Lt+EeWKk0oJrkT+c5xmeqto65qSAWqYiZiuNaa0D2A7n2Ccfg4cZbrfF60enMtMV4DSRZqlXpZDgffzhmV6rO8mKB7ikKNTfs6v0F+/Gn0Y/HZDR44GXlI6OJLwz1jt8vPGFXPsQ7eZ4Z+QE5b3/ejJy4EIgKluN1Q79YpfCyXDSEiSoGQibuAT9LVeMZsq0o4RVUr6D6xkD19ueO3wQj/RNFekCW6UPrT1UjJtBC8BADJGQ2Ahu3wApRCnvMxgVwFMhi9PAMCxUr8KU+Dnj7YVulMqDv0floToISoHkosGoJD2hvi39BMMMrtF6h9Q1B64d+vfpAkY2Ss2sf7R3/AK/z9ZBvjJEcs8saIwF46arSDLDwSGXRYPguwCuBlNeF3dDu7iD8CqpXUH1boBpt4vYMCDmUUKjxzJA2/WmKHVyokj6vlZJvSKFVBa5qySoNKQQT4vNS/M0bP3P0QQIFTZEV17wERDrxuPX2wzr6i0oWpMS/PmQOJafyenPz4V3jdrxJwRNjyFpwJk6FBItWS6ZOieeRdEeJEN0+ouuOFi8ypYXNyykqmopsc7lvTzF/qtbOCmS+3WrP0M89+WRdzjbklwmUeAeDBWHRusOhnwzwbnyJ4DWGxKXPqSf22KXPyIisVCLZqDidJlvSVZbWMnYxgIaoKbCRB9Lv9GgtCCirg1BLOscXhCFRKfx5Z9W04kL3sYSUPiKXictU++DUmamv3SrR80JkssQovPA+IkwEHOx0W3NoX6/JKVQqzp9TGn53HnWx2jDzd13CqyXJWvKHqbFQmSny8KS3UA7+PB1y+cQfRFmXTLbcHL3Ag1XUijqyyC3TPKFhcMoxjZR14XT2/Pfk+vs6t5YmsIYP1faUEnpdJchLZo3Kbor9eu69mrKufkc7yLZiAnzDtwZGj+5FPywcFugmIgU1Q0dymTBfo6zHxO6oobsG7PQ5Se5GCsyyTCRoQGqJOrS3YRTIkT4zbuN8asTfsFdEGleJvG1aso4WETV5naYrLcxMKZyG5BrFP3JAyTT5P1xykgYBWVKTii2H1gHedOsOeO+eN+m9oXJA3YauY0t7j0TwQmVvdqSl5jQJ6w6vpAxP0zMieXqhuI+XeqGy3WG9UOy5rOLxHSRbjBtCPiTCrTx7didyDIFnBXq0dSyGgKKmEmWnp1S4qdiTZJBiE4XwjBdttpyi9kJWtX0xqWJVIjnwpQX1d9TxxR1Iw7Dpn0NR0Cc2d92L/SpDz10rOOjidzXaBlHQFYzQAWEC/qrEQJ5eaaLaluDJvNBx7gr3SqUKBbo91xTnA+zRS2zd8ef8kGBhsmQXJuhgdOg+5mSgbGUXf9mQn4RBotBPdyYfgWuXNtzOZGAIcFPCHESW22kCMV88TfAPd5q50xg/TZj6WUdDhrVFBxMTcaQynXNd1tXOGrKrahHhYkUNT3C/OBw40VlvoPsGaTeZvwvJDqNfF5Ml4CC1cGWqQWI/RM2TkQmSZkBsFKy499kRu/n4rlJoWwqtl8oTFMwdE7bCxA+GSY15K+hTG565AF0XTyHifXF+MdkCyZ58rtG5GN+4ekfLw//i714219q436Zt5E5Tk1TTuebVi9rLsznqAPL++t2YEZsqFFnHN67qxiovbhLHsRsPg1mexk077Lbx3cV80uTA9crbDDOz9PQ/u0ox5XU/1TgQuLmuJ+UB3T9kaeV/fvf2zX5Ib5v/zY377ga7B/hmUu90z1Fm7SSdGkxIWMo62A+BullcLCYSLlMUs5dniwKJr7m+9yik4t6T5vvd7gnXmi/6D4Iky9vlpqGyU4V5VYYjlptVs2yl3H5jOnr51xT7qu/E15cWG1YP61PtjpgZ2rf/rXu792mKTvOto1sndQJiBw0uo9lc4Z9Rpdw3kxW3OKBGJ/envKhyfkKPjxrRTrqt0lFTFKnZcCti47xWa9ps5dZWZycnXC6OciQAyVFR8JIfGYu97DEWLLQ73YtxfLALd/FdkwM+39I72jFddduoX7vNWrdN6hLI2laoe7VpyeM3M6OVlUq77Ni8YOjWAqMtQ/3pE+P36bbJ+HR9aO3NKZ0m/Ulw1JvhyJODsas/Im2Zfzq6zbjSZPDBvOAeb2mih+1t04ZSezkatoejTS3daFsv9GR/EsTd3SiEQ7vqdjiyXlEbxZMm81Bt2p3KN2S+pVump8olwQDSjwQGdk1gYNdh6e3bXnQXIsnnofvTDTs9+o7ArGzL270uPexTtjlO2tj03ePgav6SbPAPZ0dwkzjIxFnVuG4SqQhKfNovwN0yhNptnfIYjCfKNfwlaYmttNvurxOf0X9oKACVwX4uSPWJoGlceaCsUnK3RJZYu/faWa9YpLej3oNEcKiFB3tSFRjuJJPLXI8hAX6OQgKMvLBtCnR99HoSxIxFiY4uPj7OEGQ3ulgu6TGmML3w23X31c6MIuAzVZ0ceIJli5LfHeCR6MKrcjTxUwYCt3aYGq7eqYr5G+dxDJXdefa2l+PHl9cTSiRhvU/2wKea08RIP8+iyOOcBgg84J49RgWXWU19+VnkadK/fwDqG7YT +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Run preference optimization of a model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/run-supervised-fine-tuning-of-a-model.api.mdx b/versioned_docs/version-v0.2.23/api/run-supervised-fine-tuning-of-a-model.api.mdx new file mode 100644 index 0000000..fa76549 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/run-supervised-fine-tuning-of-a-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: run-supervised-fine-tuning-of-a-model +title: "Run supervised fine-tuning of a model." +description: "Run supervised fine-tuning of a model." +sidebar_label: "Run supervised fine-tuning of a model." +hide_title: true +hide_table_of_contents: true +api: eJztWltv2zgW/iuEnhIgTpzuzs5OgX3INNPdLtI2kwsGizYQaIm22UiihqTieIL89z3nkJJoSXacTHaBAs5DbEvkufE7N5IPkRamVIURJnr7EL0Zj/EjFSbRsrRSFdHb6ISdK2OvNJeFLGb/VpPD6CBKVGFFYXE0L8tMJhxHH30zOOUhMslc5By/2WUpgIiafBOJhYmlVqXQVjqG39QkriqZBiON1cAmenw8iHiaSiTLs/Ng1pRnRhyA4L9XUguY+qUlc3MQWWkzpNMRGggiyb8OaXg1FwzJCWPZghsmizueyZQpzXKeTZXOgc2r6Wwst5UJxkkgOxMaBq6K9a+rq3PmRrNEpSJ6bLTrWqs79RetQXgafMA4M3OlLTNVnnO9ZGrKLGgsaMxiLpM5k05pLXlhGSjMeOHfE59HpG+5zLbl7EYj60wVoBubVzkvRlrwlE8ywYI5K+IgJ1mAzkWyhZZ7n0sHj312wq4vzrwyCQg/EawyImVWwcrCbHEnWK60ADVxPWnlGJ+oyhJzU4pETmXCVJJUWgtg35HrOXD0K1yvVmO8AJ1kpjXGq12STbXKSYaT8w+H7EJVs3m2hOXJMrUw7OL9O/bj38c/HqLRxD3PS4eMGl8A9ZbfzzxlFw7irTxbIt97zpufhj0nySR4BZvDdINfrFIwuVjWhJEqAGEm7wR85qqCMWhbmYudU+2c6jtzqjc/tfyuAOkfEenes0zoWv9RFfgEWEjcJ0KkaDZ0Nm4Fy2Qu7SE7zwQHgSygh88gUbEMXupD728/rMtURug7WHwwJ7qSAPMgsKpC3IO9LfwS3gw719q51nfkWj+E+eoDIhskZ5cO7S1/714nfcjXxkgP2ecKkCB4TllpIiDxFMqCweCdd69UTHmV2YFydwPhnVPtnOr7cqrxELctXIi8BKHGZwa1CbsptvdO5fh5qVSxjwqtKnBRFQAzsM2dROtPZSFGtqKZYE4OS5CKDJvIkmueC3B0ZHHjzAdp9GeVLlGH/3mP2RUcs+v19YfTetVhJmInAcRaQRa23gQxyDaVs+EEXQ9iblClSW5UeLOgRSxKlcy3iQqfqnwCDoVy1szcXMJ6VaCsOb+PjRWliYGHI72Gso+Ex10uH/m9zKucFQ03oud5sBKrECIL3Gaap9gHxBwQlFcZ6ez4P5Prpz63hqZgNR8M61P05apMgRfuGDiVqYF5OfcgnPTVb2l72VZMAG/4WmAEdN+FsKBAizMhSIJmsJC8SJkLT9aBdDNqcK4RNt4G4NeFBA9jMgUDYjbUxN2GmPXkUJ8Jt8k8NvIP8SxEGgpCzjYNWaKFRM28mk5XstdEKSiEix7F3+YCJNO4/n4SSeoFZGmFKjYcmgVwpusvwHt6Xrt2TWUPE42uEostbyp5pmb7rckbMwrAAIbBejQKS4PDcH7qaDpGKE8AxeesUgCV9QsWQDFYspIntyJdY1wPeZ+r1vIM7I7kGDielbCizcICBBTWEyA7PsWgjakSJRNTyJ/gnsmyiZYxaC+LsrKvJlWiciAHbZQyhoH+RB1e3IrCMKj3FiLL8BPyevvieck7WK4VP2jxu4q2Dgqch6MJFCiQw0z9ssCAK+1J+Golmykt7Tx/Mi60nN24XqaCp+gNzTg0LZRLbA/KsxyqNvi/OEC3MLN0k0/gwOiAPhZooNnKNuznmvyVryEzHQDBhdaeB5wJrilsULvcMYSgAnEh5Gxu41QkfPk0wd9oNKPRgJ8apq7M1WIGuUV7EyNxoBIvuM6rcmMO2ZS1kHC2ooYj+DwcdhaRrNfRfUDaIfO3kGx99GWYzAXU0EtKU7UnhhA1TyJTFFj+Q6Fg5Z2LjlDJJbelAtsitF4rTiCYWyZshYnrCdIK4pbXpzJ8RgDti6fA411yfjXZPMlAPip03p1fU77DfaN/wncnG5U29C1ukBtPTVrGC83LV7WXYzNqHeT95ek5QzalT7LENymr2iqvbhLi2LYG3ixP+03T5zT4bjGf1jGwn3m9oNFkevy3TakY4zo1ws4JsIcJpdzD+QdsWrr/f3mz/zxPb4r/4cJ9c4EdOHzdpLW6z0FmTZLGBgISpLLW7buOOiwuJJNCfJ6CmEGczTIg3lv64JEPxcGT+mivfcK15svwgZfk8eZxqDlrVWFOlW6LhflFzWYr6fY709HJP6AYQW67JtahExyn7r1d99pGwDiVejtSMBCExQBZx6WWyp7Zh6dEuqlLNnbGzah+a9yz/prksc6RgaBBNz5TFycrft08WNnkawQhj27ra6h7sgVfGvaVJn6NHKQ0j6EdK2IwbJWFgvjVPYikFbnpC9Zb4jNpqAkBcsgT8pKjyQqIJK7ZLctsyZA7/CLb4oOYhLAqzrN1EX9N87RCj308O2cZX+KeS4+0K5j/DHVfcsOS4eKhdi0vzYvbLQqpCxhWd2lEFgrL0roUvpepBbBFSuwfbCrwRxuD9x0Qyznfgg2RNgl0UVAMNDSAK/SIuPGkVWZC1rCeophZ6mEhUcYp2OxVM+8pCrTnCtXRqcDSSuH22ZlajMgmJ40wpOjvFQf8/CHiCXQqryZJTZXMj5R9LHEl5zO7qLpm7bnPEKbXgNHjpl7YINOdwcD3EN/c1qLPd10t++XrgD3ZHqJhP9ypPBwI2M8NRb+eXK1Eovr3cwMRzHNxqF4cHWOweDp+f4JRtSv5uZ0u0oPPbeApbCC22+25RIgAZZrkmh36CngNGb0ILh0tVyQLlh+s8oLV/zWQbnQCXZNg7bY2kOyCADe3JZLMZcEhDaJtPAqWn2gRarFzXyLDAEo5T+5L7zLZLpPtMtkuk/0fMhmlnj8dkHb57PvPZ4/t9YF2req9ghddue0fim7Ycui254H+l82JMZrhCnrW+r4iCd3ytroStCFm5wp+RhBcCMscQtrb6OjumPzrCB+PatGO2vPoUdMT42Y13S4x1HdWGg/F59aWb4+OeLEczYECjM8ynvORsTy5PYRYRjWBEUkFplvSxNPaI77c4Ds8n75oT7J/aQ/l25PoFuEDJ8rtOfB4zSnu8RPnrcfrjkKP+6eUwcFUK1V49DcODu3Q9p1ztvBMbM2BV0u3Pp+qt+w6B0T0eM2pSfc8oz53wPOEcfc8YDy0hz9et/n95Ia0F3fzzrAftGmj1g/pb6HWiqd16sPNyM17dwNbXe0mURC8O3s+7ZuhLRvP3leaA9Xil3r+zWA15/AxXIu5d652GjflzTgsQrx9OgWBMw5dCpoqkrLOm+iZ7BI9k136yzsuTgMUwa1d0Lw7HtjYwrtOZuXWj4updbkWUKZCCjLlHSQz87VgnT/O8BwZ5okiJTsbOhoBKhJ38bS7yEMn+5h99JQnWB0jUYjWXwu8jKM03VLqE5/gxexMgDJ8Vkv1EZfYUFrHGJdzug3jMuj2d29WDBLctdmegseKFff2qMzAkamg1LRF6YLxl8gH48iJ2oRjjLxDARlAhUEXZz484OJf6+zxER9DONVLd0uIbqhNEABfcAd1LngKGR0D8a1YuiSJqoyuXPqFmFRRuOheIcIG3804SRJR2o1jb4KEc/75Egufib+mhAaBp5ovqMRbwPfIhTA8DIMB9Owhyngxq/CM6W3kaOLffwGJpjsh +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Run supervised fine-tuning of a model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/safety.tag.mdx b/versioned_docs/version-v0.2.23/api/safety.tag.mdx new file mode 100644 index 0000000..2d509b6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/safety.tag.mdx @@ -0,0 +1,19 @@ +--- +id: safety +title: "Safety" +description: "Safety" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/save-spans-to-a-dataset.api.mdx b/versioned_docs/version-v0.2.23/api/save-spans-to-a-dataset.api.mdx new file mode 100644 index 0000000..c64bed8 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/save-spans-to-a-dataset.api.mdx @@ -0,0 +1,68 @@ +--- +id: save-spans-to-a-dataset +title: "Save spans to a dataset." +description: "Save spans to a dataset." +sidebar_label: "Save spans to a dataset." +hide_title: true +hide_table_of_contents: true +api: eJztWE1v2zgQ/SuETrtAYqdFi25zS5MuttgWySbuYZEGAS2NLbYUqZKUE8Pwf983pGTLsZO6wF4KxIdEluZ73puhvMgc+doaTz47XmQvj474X0E+d6oOyprsODv/O1seZK92PRqVJBx9b8gHcSe9UGYmtSqEdaKSemJdRUV2kOXWBDKB9WVda5VL1h9+9Wxkkfm8pEryVZjXBLN2/JXyAMXa2ZpcUCk6H2RofE9OweyUHAQ3w/prNLoQSVrktiCOP6igqafrg1NmuqX63jkEH4UPhBS+tC4I31SVdHNhJyIgY4oyd6XKS6FS0k5JEwQSFtK0z6OfJdsPUul9PSdpdq2tQW6ibCppDh3JQo41iZ7ORjjsSRnkbPI9svztPF5J/bs4EZ8vP7bJ5Ah+TKLxVIhg0Vlo04xEZR0hTe5n7JyQY9uE6NzXlKuJyoXN88Y5gvsHcSEwWRQq+bvoNXQitaeDjPGjHHByfN11uOvWqng3q/6lMj1SvA7LYuJsFWM4ufgwEJe2mZZ6jvZobe+8uPzzVLz54+jNgItG97KqEzI6fAHqa3/vZCEuE8TX8eyJ/OWSs3/18u1u5uRagRWihLrni2AtlM28M8xWAYSpmhH+V7aBDNdWVfRMqmdS/WKkevl27W8EpH9ipLfM8n1q/WsbcAIVovucqOCyMdlkIKFVpcJAXGiSCCgAPXIqQRKNh27Q8u31Y5vKk5uh+SgnU4lQHgZWY+ge9Q74Rm0Znqn1TK1fiFqv+/vqAyMbkYurhPa1/5ZeJ9uQ74pRDMR5AySQrOJWGhMWj7EBBcOzll4FTWSjwzbFnjL8TKpnUv1apDra5W0PCkWWMNTk1HM2I9JUoYZzjn8z3itecr6WxnOhpShkwFoLA8a8dBJa5NjGTaoP9uQ7W8w5yP+JSzKgt+Mm0O1E6eRspSKdk3NoqECV/7GpbzT/MeJ4Ca98Cqhw3sm1gACqZutHjsq2QkmUZ8DDqwzMLdQMyXOQD92SaSouPn3HF8MYmnLImFo9EP3TkJufWpNgeV6zfxzimwgDa+h8AhOrlEyjNSRW38fW4hRi+rfgdEyuf6cNqHcnlbV3o63ncgsdnHYMhxNN+VM67uBN5KcIxb2Jpe3ye7QIWx07gef2WRxsqVnISXxnRSYbdoEfpBn3eK9bfK16xhdpcgD9UXsl62+DvfWgxh5gXNX3KefRKxvsnK54xnrp+lYV+8H3w1k32VrNtfFyTeZoupL3twXVodxjWbFpyKuqqUTU6bwERzT4uXZvs3pndTdy7yGCp9IV5zGyZ0mge/2Nc23tKbiGOE0KpcXXrLbxFbmWnHE2nL0Yhm72DWNhhhiZ2KSQSSdxH/nVOB6pZQj18XCId4LDEnaoONRaVvIQ4zj/NgD8I0M8YciqMI+KZ91B5PqGn/Hwu1yPyffrkb5jzF23E2vdap48PC/Y1C4sXneiN5ugWVvo9fsoLuKJjZ1v6/qR8xFXnI+4ahembEnHxUhAmL3YgQ0+X/iNTSv7u79vOZQyCIzmmYKRL0Y8+OAwQ/FHBDJFbYFGj7NKfMFSDlR3aXkWTHCGqpvInPnDRjF+vhhegNbFk8G28TH/GKJxVnBy2kX1CScvnRjO+MAxhmtisNyeXoAbJegtvKd0Wo4Fug/DWmNQstcIr0WLyutU3xUuGYpsildGwia6y/hj0cViDMufnV4u+XYceGkbx6PemLsK6EEB5zJweg2q0xTt4YgDWi2V7VXNayBpnOQ5kPOk7E2PahfnVyMIj9vjQMXH2+PMyTvc5L/HWUQ0a8dxGe8tMi3NtEFv8DzZ5M9/ISCw9w== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Save spans to a dataset. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/score-a-batch-of-rows.api.mdx b/versioned_docs/version-v0.2.23/api/score-a-batch-of-rows.api.mdx new file mode 100644 index 0000000..f399118 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/score-a-batch-of-rows.api.mdx @@ -0,0 +1,68 @@ +--- +id: score-a-batch-of-rows +title: "Score a batch of rows." +description: "Score a batch of rows." +sidebar_label: "Score a batch of rows." +hide_title: true +hide_table_of_contents: true +api: eJztWm1v2zYQ/iuEPm1A7KTFiq7+5vQFy9CsRpJiGNLAOEu0xFYiNZJKYgT+77sjKUu25Cbuy9oAzodEkcjjvTzPkTzyLtLclEoabqLRXfT06Ij+JNzEWpRWKBmNojE7j5Xmx2Dj7Cy0HkYHUayk5dJSByjLXMRAHQ4/Gup1F5k44wXQk12UHOWo2UceW+xYalVybYUfMwELhtupSFptjdVCpth2XZVf3rknyH9lFxlnIsHxxVxwzdScWXwThOEzWHYDhhlSPYmWB2RolVvTpxAkifByJ2uqfV5xJ3qq1U27LWgNC2wqLC92GktJ/m4ejS5XPWSV56j36v+ZUjkH2X4lq2LGdftN8FvrjVeo9SJosrxaLjfdSz4lq1AE8+5ic6UZhzhjaOeQvQ5PTBgGrICS/B6rvCokk1BwZhW7hrxCfKBwSFPNU7A8mX6B83+AQzb9ceoNLDgKiVcGNmZ5WyNyZL8Vc8gNJ+j9WwmC4eiyjZpeD10dRFbYnBQ796E4c186XBj3hQqYwTc5d+GKOgaN65jVXeeVjOnbyrj1MXeybIsF64mjY0b9gc21KtiM2q60o+FcUjEMVQzcNmjXkhT7rS9ZEYRJKW48/4XEGImEoW8KyNFFBSr7zXKXsWCrNqoFik0RgZtW/nFxMWG+NfIl4RSa4KT7ct5rrVF51/iA4pspbZmpigL0ok573LW5yQQ6T3ijtQAZMCHDdzeOw4QFkT90ZN+ahs6VRNtYVhUgB5pDAjNEWqvPmjo0kpBos4wfYGUrs4/Z+7O3wZgYlZ9xVhkkG6JTExP5NWcF4grNpHi6yDGYqcq6wU3JY5wTYqbiuNKa4/Abeu1EVx/hOlor57VA7ty0xXl6Dd2kw3hyMmRnqkqzfIHhyXPMBOzszUv2/Pej546y/BaK0iOjxhdCvRnvGBJ25iHe6PNA5AfmPH3Rz5w4F8gKltHESQ9WKewsF7VgkopASMU1x7+FqrAN+VYUfE+qPakeGamevmjGu0CknxLSA7NMm1r/qAo5gR7itzHnCbmNyIbzNstFIeyQTXAlggpZRA+kgCTJ8aOuZ6pn22Yqw/U1Bh/dSVTi6B4CViX5LfqbFhg8uGFPrT21HhG1nrXnqxNCNmrOzj3am/EDvcZdyNfOSIbsXYVI4FC4WWnGceKRym36kppeCZ8DrVa7O9fPCN6Tak+qx0Wqo77RHkAhxxKCGqSGrAlbPNJ+XVu3W8MQ+00Yuot2qVTpKUHj9hBZTP2vvG9wjjxWyYIU/AGFIJo8T151qj7K13ucB8Mmclpvcb+w+NApQ2zR3n91zsB1+SjK82IKZvqxSlKPmJCjuh827SJJvXv0JgxIxPwGFoZtCNt0GpdVQSHfaKYxPd1OUZxxWWoGRsQ9hYc3ckIjmouQMlz3aYGZ6wF546RTm3v79pS5zhQnJDMDKiI5oS5FBXtpJHRtUdqp5cgBXEjtlDleVsYizbwIVotwI5AWfjxvxMqmUJQht/CHVPNWVaVOMcNJIAP5rdUQ+2wE0tygHxz9Uy6pqIEsrdNCu1SGYnoBe48iTaRxlarBBfmGizSjylLzCneAOCHQzIcKpOjtGPKpW3oSFShdQLxoA2HcqPUmaOXA0IEtvjQU55YdK9yadnRDscr0Fad6+1LBDZPKgh5sqE/6seqy5G653Hr921juR8G2oLTcg4gemz+p7wZjOg6arKjrnIEdB2AGHosdnmMKmYu08qWvYU+p8r7cs8HvJvd0PnxN7tkQ9n1zD3WgbP6NOLpn6E/P0M2IP4CPLrQTB60dCekGYR6V34GQHuVtJjZvvoaCtZTvy7099P9n6N8L9GOK3I4Qd9G+H9u0KxAkphASLO4yMJ4B3Iu/UB4KDloWaDrFFxus4ezehf5+qbxfKu+Xyo8mGz2qpfI6YXdORfuV837l/OgJ+zOvnANsdybmfgW9p8AjWUEvl1vxsH5Xja6zdS+dtS/hrfmJ1oH10iw08IVuREB9fWwaSuEtDNXX5DbH+TvjKEm7ojmdbbsjbS/FRaWuqg93C0Crit9Xg9+i7pY7Y/6Si/NoM4bVFV8S2m2m8N+oVO4iTAk2w/8Or58chmHdXz5wxxk0sDt+NK6WX2k6NcmsLUeHhyAXgwyF8GSQ51DAwFiIPw1jVbjtkOHIHGEXruOrOv9cXtE3OuM4a05DXjenNu3TjIbHfYcSW0MYDKUTLhfPeilFOrJz0pGdh3Muhz2UTwb68F4/6cmNdCxo1g7IPB/rbUdLsrtBizn5WqCQD5Jt/ACjAxeiqUxKJSSCBqS7FyE0MkL7BUBCIKbjUD2HmBb7JBQ0/yDp3Ipu5iLUusJndIcp5z6xBa1Oaf3p8woFvAB3tiT9jnTrydWaA1onVdt7BNZYXOMc4nZI0P1SD5e7ALFL79t6D+afapghJghK1OruDhMGf6/z5ZJeI0j0wp+fuYPZGQXzkvifcUhosYrw+sSxSfTSqzkIydjfMh11D9com/ge4zjmpf1s26sWZSbvzi9o+goHeLS2pzUu3NAUh79HET6ocjVNuXe4zQeZVjTXjCIvk37+A07qa5o= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Score a batch of rows. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/score-a-list-of-rows.api.mdx b/versioned_docs/version-v0.2.23/api/score-a-list-of-rows.api.mdx new file mode 100644 index 0000000..75b2f35 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/score-a-list-of-rows.api.mdx @@ -0,0 +1,68 @@ +--- +id: score-a-list-of-rows +title: "Score a list of rows." +description: "Score a list of rows." +sidebar_label: "Score a list of rows." +hide_title: true +hide_table_of_contents: true +api: eJztWltv2zYU/iuEnjYgcdJixVa/uV2LdWgWw0kfhjQwaOnYYiuRKkklMQz/951DUpYsy03cy9YMykOsC3l4Lt93eEhqFWkwhZIGTDRcRU9PT+knARNrUVihZDSMRuwiVhomoSFTsw8QWxYrabmQQi6YVreGcZkwvlhoWHALCUO5ZWbNIDqKqCVIS5J5UWQi5iT55IMh8avIxCnknK7ssgAc0A+AHQutCtBWeOWCxK6GPEkEyeTZeKvL5wUaMmtKyjfacq35EpsKC/lBYykJ5/NoeLXpIcssi9ZHm/uZUhlw2Xwky3wGuvnEWI0ebT7xCjUeBE3W1+v1UStUlykwssoFxbmLzZVmwOOUgjRgr8IVExgwlvOCqTlGMitzySTPgVnFbnhWwgBHjOpwTr/A+f+BQ9r+OPMG5oBC4o2BDZQ6WyNyZLcVc54ZOELofSqFhgStaaKm00PXR5EVNiPFLnwoJu5N1FZu1BUqzgw+ycCFK9oxaFTFrOo6L2VM7zbGbY05OMy0PSZsqL9jAuGtSiBsrlVe6UUD09C/dCUU3+tTCcayW26YkBgGkTA0P+cZeiFHdb5Z2jCW27IJXIFiFwiytjF/XF6OmW+NlEiAvB/c0EZju+srrVF51/iIQpgqbZkp85zrJUXLosXg2tymAhkovNFacBnCLsN7N44LOybX7KEj+9Y0dKYk2sbSMufyWANP+AzB1OizpQ6NJCTaLOMHWPnTeeFh9DMbsXeTt8GYGJWfASsN8gkBqIlscAMsR+SgmRRPFznGZ6q0bnBTQCzmSEkVx6XWgMO39DqIkT7CVbQ2zmvA2Llpj/O2AUw6jMZvBmyiykWaLTE8WUbz2+T1S/brb6e/OlbCHc8Lj4wKXwj1erwXPGETD/FanwciPzDn6fNu5sSZQFawFLsburBKYWe5rASTVATCQtwA/uaqxDbkW5FDT6qeVI+MVE+f1+NdItLPCOmBWaZJrb9ViZxAD8FdDJCQ24hsODWzTOTCDtgYiw1UyCJ6+AIrV5bhS13NVM/2zVQG9A0GH91JVAJ0DwGrlHCH/qYaAoIbemr11HpE1HrWnK/eELJRc3bh0V6PH+g12oV85YxkwM5LRALw3M1KM8CJRyqLDsN3gV4JzDkVwbury88I7knVk+pxkeq0a7QHUMixhKDGF4asCas40n5bW7ceoxAL48o6WofSHkvBNS4AkcTU/dq7BqfIFypZkn7fiEZCFqX93+2YuO0rBK1b2bvIhlXstFpdf6FZOwbucat/66KE64VhlGX5lJvphzJZeCSH3Ln7om0KSercHqjxgQkiu+VLw1rC2rQHdDVBsdVMY9q8m6I447LnjBsRd+x5vJZjGtFchlTmuk9zzKgPyGdvEsQqTR+6Sghv354x15nihEmGcdq/ckJd6gz20kjo2rywUwRjQQXeQRntZWks0t+LYJUINwJp4cfzRmxsCvtB5BZ4CC02eG2pMfESyEC4s5rHPktyaW7RDy4tLUCCrnZX/Y5MY5cOxXQC9h5F6khj9ay5C/ItiEVKm1r1I1yZ4kRFMzIqsEBvxzybupKYqEBpjMfLJhBGtVqvg1YODDuwxYeG4tywY4Nb04xuvae8uy/W2Zf2+jDbLenChq1RP1a1I3rYHGO9/k0sd6NgX1Aa7kFEj8yf1LfFmB0HjTfUdc7AjsfcHHss7vAcU8hcLDAYdDfoSIL35Z4Wv+vcs/Pia3JPS9j3zT3UgbL5N+Joz9AfnqHtiD+Ajy60YwetAwnpBmEeld+BkB7lTSbWT76GgpWU78u9Hvr/MvTvBfoLityBEHfRvh/btFoRJCYXkltc/WA8A7iXf6E8FBy0zNF0ii822MLZvSuQvlTuS+W+VH402ehRlcrbhD04FfWVc185P3rC/siVc4DtwcTsK+ieAo+kgl6v9+JhexecNsp3v3drfv+35SeqA6vSrPGJ1gHOaWz9d+2P734y5j9+cRbVcqwuYU1os6nC26hQ7gOZgtsU705unpwE0e6XouGP343bRS81naOk1hbDkxMul8cpdofkOMt4zo+N5fHHQaxytxAxgJgVduk6/l4x/+qa3tGxx6Q+IHlVn+M0DziuVk5Ox1HA2p2YzZVjW1WkkA7sgnRgF+Fky0UVbSADfIBunnRkHToINFtHYh7pVUHfkGxTbqnmvhEo5L1krT/ODLhjIZBJoYS0/rNclCI0Yk37qTUheNABqJ7zmMpoEso1vKcPejOl3RHervAZfbWUgU8ZQaszquw8YymUOXfHSdKv9fadVW3Z3zib2tshoN5i7XCCywxBJ0IeDKsAnSvv2Wpt46+AUEkQoferFVIQ3ulsvabHGHy99Edl7gh2RkG8IkalwBMq/xA2HwGbRC+9fschvflPRoe752jET99jFMdQ2M+2vW6QYHx+cUkTQjiro2qZqkZ+S5MG/h9GeKGKTeJ3z3DhzOWipOw9jLxM+vsH2HlBig== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Score a list of rows. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/scoring-functions.tag.mdx b/versioned_docs/version-v0.2.23/api/scoring-functions.tag.mdx new file mode 100644 index 0000000..38391ac --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/scoring-functions.tag.mdx @@ -0,0 +1,19 @@ +--- +id: scoring-functions +title: "ScoringFunctions" +description: "ScoringFunctions" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/scoring.tag.mdx b/versioned_docs/version-v0.2.23/api/scoring.tag.mdx new file mode 100644 index 0000000..fa00830 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/scoring.tag.mdx @@ -0,0 +1,19 @@ +--- +id: scoring +title: "Scoring" +description: "Scoring" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/search-for-chunks-in-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/search-for-chunks-in-a-vector-store.api.mdx new file mode 100644 index 0000000..4d4295d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/search-for-chunks-in-a-vector-store.api.mdx @@ -0,0 +1,69 @@ +--- +id: search-for-chunks-in-a-vector-store +title: "Search for chunks in a vector store." +description: "Search for chunks in a vector store." +sidebar_label: "Search for chunks in a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztWVtP4zgU/itWngap0DLa0ezwxty0aIcFlc6uVhRVbuI2Hhw7YzstFep/33Nsp3HaAGV3HnYkkICmsc/9+45zcp9oZkolDTPJyX3yejDAfxkzqeal5UomJ8kp+ZOlVukr+GVXjOo0H4ZNJFXSUi65nBObM2LcXQIyK2HNUdJLcAGTFqXSshQ8pSi1/82g6PvEpDkrKH6yq5KBMjX9BspgY6lVybTl3rDwdbPOWA1KYV3GZhSUwVcLZ+XEoJlH3pJJbUlJ58wtjh27cEIJSiQ8Ayv5jDNNZkp3OEOciHUvCZK/V0yvug2KdYxAkNJ8ziUVtUS3FVRQS5bUEHbH0sqyDIVn1MbRoFrTFcjklhXm6SjNuGATnj1t1VfJwYjYaTVzPqOE7aT6AKB1eFfSgj2t4A9YtZ9Mk0K6IoGyKqZM7wgcMsEWVKaQFtwQksRNO0sokFowaQoB7QwYzTKOIqm4bBeYZBez5OR627F1b9uy6JupUoJRmaxv1tv2vroovZoD8jtbHS6ogHg3phFqjEo5hbSTJbf5JlDoQYSZf1cH/u5ujkCwQaRYdmd3IvzBa3Vw6JG00hquxIooCX/Gbs84IRjxqiyVDgXrRO2FApraCjCAG0jt4Xr9UEZmVBjWA3b6XnENqk6uvYqg8Qb+cytQZUROwYUHXcPwkZlWBaHEswVxbOFLFK/iYjpKdrL6hRuLZZ1GAg0pqE3zLQr09PAs/2r0Rjir4dHURLfjbVbuAE+ga+e7txDtbUfhEX9bcSG+6gwuz6mZFG0A16CIqDk425b9V84gXo5qIQMUflHQhm7pgnJBp5CXKVspmXm01yQsoQgm7uLJ2ouQOFK3TDrq0AwWs0WdNBTnZKOzwYLnJW+DxFZ7CIQexWmf/F12NSv4EnoI8oXeN5tr9OCXrpY+chQMNhrfgrgEguIZQqCgAgJUsOzHNW9jqa1i2uIgdt5B8r+NRpfErwaIZS7TIVpPZfmT1tgRcHEPImFyYChgqqKg0GpDI2JuzTLnUMrcO605BRhjRVAZ7js9DgnQsMS+mv1qVC2UBN9IXhVUHmpGM1fE0Z6WOaiJAytja3tWLZ+Sr8MvwZkUjJ8yUhmoDqvq4g6A4hLz6TJH6FRV1tNUyVLo/ClRqaf6lG3Z9ZzqDxmus7UJXlTtLkwPBK9d0WjD6eXZERmqap5D85kpIdTSkOHnD+Ttr4O3jqjYHS1KXxl1fUGpN/re04wMfYk39uxZ+QE5r991IycVHMkfMA3ow46pFGyWq1owSoVCmPMFsA0tVCUdiVpesBdQvYDqJwPV63eNvhFU+jlWekCWiaH1t6oAExAhdpcylmHYEGzQs4jgBbdH5BIOBmCQheqhc3gaIAJu6rpTvXmoUxmmF5B8CCdCCU4LGRZWBU0b4o0dkYUwvEDrBVo/EbTexP3qDCsbhwRXvtob/QFep7slXwcjOyIXFVQCo4XrSlMGjUcq92if1fDanMZ35juPCH4B1Quofi5QDbq07QEhhxIsNTo36I1/Oju7QPPb5vqHNVcvaV7J23Dcaz18jaVfhZOerVGDc9UNs2y9f0oxaZiaMBuk8LyrQpb9cGIzOsJLYAo3XC2ppgXDKze58qO51hzUzxM4ml1Sm3cOZs4+1klvGQo15J8tUVOTZ6sr1ttFfT0uW9/4xXA4eK8yNyD9QQSyGbjuMap7aF4WW7kbCB97vwYfCJwUlzAwA/HTnvEchZGoj/9/GDXKSojOseJjs8eHnI6+CJZ0TCc/e6ubwmvXmMHkS5Cnlp1jfZBX0LsJGFUP2Pfg9XN6x4uqIN6VaNAS2KrSkrw6xos3A0BUKioDD1A9EvoWOR4cOM2aylvwfOLxscc0FDeAPc/h1Xh8HfQRKuZKc5sXaCKw7GZ4PbE5eJIrEQ/eozF2aLuDR/Sdc+lio7em3BvJgTb2HExtTf+CAyFgTtSMS3Zoqwdf3GCc2RLcZTsvOuL5Xuc8D7PptvrRGtCyBhYTVM4rnLB5lKERNd143a/qRDsnDqJXLQWeAvaaMQdRuCFkiRyScXLLVkuls3HSgwuvFT+D7nGSr6aaw61Gf7Pk4HkN0EcqancXJZOU+04QTfvqsYTrN21mRWRBGFWGhK3c6MLx9knSXxz3lZOHn2KON/37Lcpf930g3DgST5O+P1Qam2BubXnS78NT3GEOGlh2KAQt6CE00PT2KFVFguxoGLRFbldu48e6hq9v8B4y97Dh+E9NEw610qSooccuzhh0ornGayNlB2aD3foMfSmumIYjEzz/uBIKmfmCLpMrdJlchVOQa0ugDePlS2px3FFleGg0reMTjQ90sWT3ig/IaMFByFiSrR84oTI3GWIyKxXwpnGNH6RwOFoo7U9EGWIUSVXPaIrUjEKpZmN8oSaUdse9XeFTnHABmzCNoPNWnUNYhEc3FhecTTEm4diwz7FmOxxRc/8fHYsCT+CLon4pKMdO6ov/PqDp2qfW4wlTfuzyHmEKrk92D1IBVoABhA7Kub9HK79qsV7j16EYr+Gje66YYrUBgjJu8HPWTdFxGF8NAyEckH0PZ53+1ocCiecg9/oRruAjcGHHGdGdEXJ4TkHggb1+VXhvdjjyb95qKTunODxz+B2nacpK++jam4jiLi+uRrB4Gk6KAbaaLvHICX+dxS1qWOK/upfAfS8Tf/4B2c2P4g== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Search for chunks in a vector store. +Searches a vector store for relevant chunks based on a query and optional file attribute filters. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/set-which-version-of-a-prompt-should-be-the-default-in-get-prompt-latest.api.mdx b/versioned_docs/version-v0.2.23/api/set-which-version-of-a-prompt-should-be-the-default-in-get-prompt-latest.api.mdx new file mode 100644 index 0000000..a836b65 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/set-which-version-of-a-prompt-should-be-the-default-in-get-prompt-latest.api.mdx @@ -0,0 +1,68 @@ +--- +id: set-which-version-of-a-prompt-should-be-the-default-in-get-prompt-latest +title: "Set which version of a prompt should be the default in get_prompt (latest)." +description: "Set which version of a prompt should be the default in get_prompt (latest)." +sidebar_label: "Set which version of a prompt should be the default in get_prompt (latest)." +hide_title: true +hide_table_of_contents: true +api: eJztWGtr3DgU/SvCX5pAJpOWlrZhWUgfyxZaEpK0sLShaOw7Y7Wy5EryJMMw/73nyvKM55FtdsmXQvIhcWzpPs495+oxzxz52hpPPjueZ0+OjvhPQT53qg7Kmuw4uyxJ1M5WdRDXKpQi4H9fU67GigoxJecxThh7LTwFIb0oaCwbHQ6zgyy3JpAJbFTWtVa5ZKPDb54tzzOfl1RJfgqzmuDLjr5RHjARDmtyQbVxte5743xwykwwbjtUP/OBqi7iQDcp7Kl0So40ctEyp9LqApEfik/ptRfSkbBGz4Rv6tq6gOSuSzKi8XAVsz7vsBInZ+8Os8VBlrLvRaaQ8ITcVmifEk57aYDwQSI/WJZBPD4QyuSOKmAFtxjm5ZT22UObx1dV/Dr7j0b9aEioAla4OE6MratkYJMoy6OaDf3x9MWgUBMVBqX05Z+PYhYdBj0f0jk5gwsFMP2278Wm8/fKB2HHHe5LtI2sgFcokWUujRgR8EQ8ykREl1WqUJVAbFb5r4lAPa8jazVJE1NO38ZSe9qM4lU7DuaLyDXAixrCk4M75ZdsVT66T8aWr4FXOy4RboF4ZFEoti71WY+TybujH41yhOJ8XnKhX7M+tmupXR1kQQXNuZ21vjZTOenAgUJt43LCQ41nLi6zBgSy8CxOazIn78RrjEXCkeDroDLW77WspLgIMv8O3i44r6e3aZ1zIhTzGpxRZiq1AiOdqKRmOlFxf7KGBELj7yCevy8vz0Q7WuS2iDxJ8P1KFG+d46Ly4AMGrYSyofCqkm7GfGUaUBxzXaq8ZGZw0iiaCZEPYFP7PfqJvA9S6bt6bkeza20N675sKmkGjmQR9dGbsxZOlIJBzia/Q5Z7p3XL0X1xIj6ev0/J9BUXLCqL2TQlUYE5SLNtD+xZjmwT+p09FzbPG+cI7jfi+i+aSBXuqrUEr0f/CNMt4HWLkxiD0jEGbrzi3DaTEo16bLW2116c//VaPH9x9Dy2ZLqR4D31+QWqr/y9kgUaeaT4Kp47Mj8p58nL3crJtYIqBBqrYJkCcovJZtYZZqsgwkRNsazIyjYmtsygKnoQ1YOofjNRPXm58ncJpn9gpidl+b60/rENNAGE6CYnKhg2FhsvTVpVKhyKMyzaCCiAPXIiIRJet1y3Uj27baXy5LDoCsDJUiJeDVGbxtAN8OZdDyUYHqT1IK3fSFrP+uvVO2Y2IhcXLdtX/pO8TrYp34FRHIrTBkwgWcVVaUTEJ7V4Oig6efV22xs70H8x/CCqB1H9XqI62uXtDhKKKmGqyYnnbNrjmufo16O9oJAQ746TgEx2pzFwrtEFF6J/8MRSN6HwNY3Z42XPh32+Oamlw8kZ0men84yP0fDRP1kqdoozX7nzGqR3C5Aq185l26viBNegWFtS7Q75i6t2MIJ6ZYsZj7gn1d/94oST6QAFfTfvmP7XEb3HPFTtTWsr3dB0Z4NY9HWk8AIlKW3BwNt4foj4H2fD6eNhi68fzpdFWgwR7SCFOljdD7T7lrawjWMCliHUx8MhdlCDEoapGGg+sQ98PLHntsq4FJ5ASRVmceKbrm1/vuJvXIDzVanergSwhPoodpOxjaCn9HsXA+IiqT5WNOtdbmXTxzvKwk3Sr7UL2W9gfcvx+gewTMFK/8WIjR8Zq4p5ZIragggeDTfuEpVDq3dtByj42oNZ4sYy7+6UpKMvhlUc70KC3TY+4hOdRsNzctJF9QHLh/axU3Ad0YsZk6Sx+9XxGmo9+dyzmyQivu4c1hpbaM4tcmueOPq5rWJiKZ6O+81kF1HBKiYjT53PR9igf3R6seDXoJljFl6t7rciJwvl+bnYfTvXz37vPClrX/yqW+1MrbulNLN4x6Yb/g+P32m21iYXVxhcYp1FZ+EI2++v2zgGl2xlNX+roS0OuhkneU7xku72sVe97nB2enGJwaPUNCverhxnTl5z98XvGKuNuMR2GN/NMy3NpAFN8b21yT8/ATvkTQA= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Set which version of a prompt should be the default in get_prompt (latest). + + + + + + + + + + + + +'"},"variables":{"type":"array","items":{"type":"string"},"description":"List of prompt variable names that can be used in the prompt template"},"is_default":{"type":"boolean","default":false,"description":"Boolean indicating whether this version is the default version for this prompt"}},"additionalProperties":false,"required":["version","prompt_id","variables","is_default"],"title":"Prompt","description":"A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/shields.tag.mdx b/versioned_docs/version-v0.2.23/api/shields.tag.mdx new file mode 100644 index 0000000..4ca1df4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/shields.tag.mdx @@ -0,0 +1,19 @@ +--- +id: shields +title: "Shields" +description: "Shields" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/sidebar.ts b/versioned_docs/version-v0.2.23/api/sidebar.ts new file mode 100644 index 0000000..43910b7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/sidebar.ts @@ -0,0 +1,1052 @@ +import type { SidebarsConfig } from "@docusaurus/plugin-content-docs"; + +const sidebar: SidebarsConfig = { + apisidebar: [ + { + type: "doc", + id: "api/llama-stack-specification", + }, + { + type: "category", + label: "Agents API for creating and interacting with agentic systems.", + link: { + type: "doc", + id: "api/agents", + }, + items: [ + { + type: "doc", + id: "api/list-all-agents", + label: "List all agents.", + className: "api-method get", + }, + { + type: "doc", + id: "api/create-an-agent-with-the-given-configuration", + label: "Create an agent with the given configuration.", + className: "api-method post", + }, + { + type: "doc", + id: "api/create-a-new-session-for-an-agent", + label: "Create a new session for an agent.", + className: "api-method post", + }, + { + type: "doc", + id: "api/create-a-new-turn-for-an-agent", + label: "Create a new turn for an agent.", + className: "api-method post", + }, + { + type: "doc", + id: "api/list-all-open-ai-responses", + label: "List all OpenAI responses.", + className: "api-method get", + }, + { + type: "doc", + id: "api/create-a-new-open-ai-response", + label: "Create a new OpenAI response.", + className: "api-method post", + }, + { + type: "doc", + id: "api/describe-an-agent-by-its-id", + label: "Describe an agent by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/delete-an-agent-by-its-id-and-its-associated-sessions-and-turns", + label: "Delete an agent by its ID and its associated sessions and turns.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/retrieve-an-agent-session-by-its-id", + label: "Retrieve an agent session by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/delete-an-agent-session-by-its-id-and-its-associated-turns", + label: "Delete an agent session by its ID and its associated turns.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/retrieve-an-open-ai-response-by-its-id", + label: "Retrieve an OpenAI response by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/delete-an-open-ai-response-by-its-id", + label: "Delete an OpenAI response by its ID.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/retrieve-an-agent-step-by-its-id", + label: "Retrieve an agent step by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/retrieve-an-agent-turn-by-its-id", + label: "Retrieve an agent turn by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/list-all-session-s-of-a-given-agent", + label: "List all session(s) of a given agent.", + className: "api-method get", + }, + { + type: "doc", + id: "api/list-input-items-for-a-given-open-ai-response", + label: "List input items for a given OpenAI response.", + className: "api-method get", + }, + { + type: "doc", + id: "api/resume-an-agent-turn-with-executed-tool-call-responses", + label: "Resume an agent turn with executed tool call responses.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Batch inference API for generating completions and chat completions.", + link: { + type: "doc", + id: "api/batch-inference-coming-soon", + }, + items: [ + { + type: "doc", + id: "api/generate-a-chat-completion-for-the-given-messages-using-the-specified-model", + label: "Generate a chat completion for the given messages using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/generate-a-completion-for-the-given-content-using-the-specified-model", + label: "Generate a completion for the given content using the specified model.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Benchmarks", + link: { + type: "doc", + id: "api/benchmarks", + }, + items: [ + { + type: "doc", + id: "api/get-a-benchmark-by-its-id", + label: "Get a benchmark by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-benchmark", + label: "Unregister a benchmark.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/get-a-benchmark-by-its-id", + label: "Get a benchmark by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-benchmark", + label: "Unregister a benchmark.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-benchmarks", + label: "List all benchmarks.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-benchmark", + label: "Register a benchmark.", + className: "api-method post", + }, + { + type: "doc", + id: "api/list-all-benchmarks", + label: "List all benchmarks.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-benchmark", + label: "Register a benchmark.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "DatasetIO", + link: { + type: "doc", + id: "api/dataset-io", + }, + items: [ + { + type: "doc", + id: "api/append-rows-to-a-dataset", + label: "Append rows to a dataset.", + className: "api-method post", + }, + { + type: "doc", + id: "api/get-a-paginated-list-of-rows-from-a-dataset", + label: "Get a paginated list of rows from a dataset.", + className: "api-method get", + }, + ], + }, + { + type: "category", + label: "Datasets", + link: { + type: "doc", + id: "api/datasets", + }, + items: [ + { + type: "doc", + id: "api/get-a-dataset-by-its-id", + label: "Get a dataset by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-dataset-by-its-id", + label: "Unregister a dataset by its ID.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-datasets", + label: "List all datasets.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-new-dataset", + label: "Register a new dataset.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Llama Stack Evaluation API for running evaluations on model and agent candidates.", + link: { + type: "doc", + id: "api/eval", + }, + items: [ + { + type: "doc", + id: "api/evaluate-a-list-of-rows-on-a-benchmark", + label: "Evaluate a list of rows on a benchmark.", + className: "api-method post", + }, + { + type: "doc", + id: "api/evaluate-a-list-of-rows-on-a-benchmark", + label: "Evaluate a list of rows on a benchmark.", + className: "api-method post", + }, + { + type: "doc", + id: "api/get-the-status-of-a-job", + label: "Get the status of a job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/cancel-a-job", + label: "Cancel a job.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/get-the-status-of-a-job", + label: "Get the status of a job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/cancel-a-job", + label: "Cancel a job.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/get-the-result-of-a-job", + label: "Get the result of a job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-the-result-of-a-job", + label: "Get the result of a job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/run-an-evaluation-on-a-benchmark", + label: "Run an evaluation on a benchmark.", + className: "api-method post", + }, + { + type: "doc", + id: "api/run-an-evaluation-on-a-benchmark", + label: "Run an evaluation on a benchmark.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Files", + link: { + type: "doc", + id: "api/files", + }, + items: [ + { + type: "doc", + id: "api/returns-information-about-a-specific-file", + label: "Returns information about a specific file.", + className: "api-method get", + }, + { + type: "doc", + id: "api/delete-a-file", + label: "Delete a file.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/returns-a-list-of-files-that-belong-to-the-users-organization", + label: "Returns a list of files that belong to the user's organization.", + className: "api-method get", + }, + { + type: "doc", + id: "api/upload-a-file-that-can-be-used-across-various-endpoints", + label: "Upload a file that can be used across various endpoints.", + className: "api-method post", + }, + { + type: "doc", + id: "api/returns-the-contents-of-the-specified-file", + label: "Returns the contents of the specified file.", + className: "api-method get", + }, + ], + }, + { + type: "category", + label: "Llama Stack Inference API for generating completions, chat completions, and embeddings.", + link: { + type: "doc", + id: "api/inference", + }, + items: [ + { + type: "doc", + id: "api/generate-chat-completions-for-a-batch-of-messages-using-the-specified-model", + label: "Generate chat completions for a batch of messages using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/generate-completions-for-a-batch-of-content-using-the-specified-model", + label: "Generate completions for a batch of content using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/generate-embeddings-for-content-pieces-using-the-specified-model", + label: "Generate embeddings for content pieces using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/describe-a-chat-completion-by-its-id", + label: "Describe a chat completion by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/list-all-chat-completions", + label: "List all chat completions.", + className: "api-method get", + }, + { + type: "doc", + id: "api/generate-an-open-ai-compatible-chat-completion-for-the-given-messages-using-the-specified-model", + label: "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/generate-an-open-ai-compatible-completion-for-the-given-prompt-using-the-specified-model", + label: "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/generate-open-ai-compatible-embeddings-for-the-given-input-using-the-specified-model", + label: "Generate OpenAI-compatible embeddings for the given input using the specified model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/rerank-a-list-of-documents-based-on-their-relevance-to-a-query", + label: "Rerank a list of documents based on their relevance to a query.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Inspect", + link: { + type: "doc", + id: "api/inspect", + }, + items: [ + { + type: "doc", + id: "api/get-the-current-health-status-of-the-service", + label: "Get the current health status of the service.", + className: "api-method get", + }, + { + type: "doc", + id: "api/list-all-available-api-routes-with-their-methods-and-implementing-providers", + label: "List all available API routes with their methods and implementing providers.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-the-version-of-the-service", + label: "Get the version of the service.", + className: "api-method get", + }, + ], + }, + { + type: "category", + label: "Models", + link: { + type: "doc", + id: "api/models", + }, + items: [ + { + type: "doc", + id: "api/get-a-model-by-its-identifier", + label: "Get a model by its identifier.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-model", + label: "Unregister a model.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-models", + label: "List all models.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-model", + label: "Register a model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/list-models-using-the-open-ai-api", + label: "List models using the OpenAI API.", + className: "api-method get", + }, + ], + }, + { + type: "category", + label: "PostTraining (Coming Soon)", + link: { + type: "doc", + id: "api/post-training-coming-soon", + }, + items: [ + { + type: "doc", + id: "api/cancel-a-training-job", + label: "Cancel a training job.", + className: "api-method post", + }, + { + type: "doc", + id: "api/cancel-a-training-job", + label: "Cancel a training job.", + className: "api-method post", + }, + { + type: "doc", + id: "api/get-the-artifacts-of-a-training-job", + label: "Get the artifacts of a training job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-the-artifacts-of-a-training-job", + label: "Get the artifacts of a training job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-the-status-of-a-training-job", + label: "Get the status of a training job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-the-status-of-a-training-job", + label: "Get the status of a training job.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-all-training-jobs", + label: "Get all training jobs.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-all-training-jobs", + label: "Get all training jobs.", + className: "api-method get", + }, + { + type: "doc", + id: "api/run-preference-optimization-of-a-model", + label: "Run preference optimization of a model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/run-preference-optimization-of-a-model", + label: "Run preference optimization of a model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/run-supervised-fine-tuning-of-a-model", + label: "Run supervised fine-tuning of a model.", + className: "api-method post", + }, + { + type: "doc", + id: "api/run-supervised-fine-tuning-of-a-model", + label: "Run supervised fine-tuning of a model.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Protocol for prompt management operations.", + link: { + type: "doc", + id: "api/prompts", + }, + items: [ + { + type: "doc", + id: "api/list-all-prompts", + label: "List all prompts.", + className: "api-method get", + }, + { + type: "doc", + id: "api/create-a-new-prompt", + label: "Create a new prompt.", + className: "api-method post", + }, + { + type: "doc", + id: "api/get-a-prompt-by-its-identifier-and-optional-version", + label: "Get a prompt by its identifier and optional version.", + className: "api-method get", + }, + { + type: "doc", + id: "api/update-an-existing-prompt-increments-version", + label: "Update an existing prompt (increments version).", + className: "api-method post", + }, + { + type: "doc", + id: "api/delete-a-prompt", + label: "Delete a prompt.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-versions-of-a-specific-prompt", + label: "List all versions of a specific prompt.", + className: "api-method get", + }, + { + type: "doc", + id: "api/set-which-version-of-a-prompt-should-be-the-default-in-get-prompt-latest", + label: "Set which version of a prompt should be the default in get_prompt (latest).", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Providers API for inspecting, listing, and modifying providers and their configurations.", + link: { + type: "doc", + id: "api/providers", + }, + items: [ + { + type: "doc", + id: "api/get-detailed-information-about-a-specific-provider", + label: "Get detailed information about a specific provider.", + className: "api-method get", + }, + { + type: "doc", + id: "api/list-all-available-providers", + label: "List all available providers.", + className: "api-method get", + }, + ], + }, + { + type: "category", + label: "Safety", + link: { + type: "doc", + id: "api/safety", + }, + items: [ + { + type: "doc", + id: "api/classifies-if-text-and-or-image-inputs-are-potentially-harmful", + label: "Classifies if text and/or image inputs are potentially harmful.", + className: "api-method post", + }, + { + type: "doc", + id: "api/run-a-shield", + label: "Run a shield.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Scoring", + link: { + type: "doc", + id: "api/scoring", + }, + items: [ + { + type: "doc", + id: "api/score-a-list-of-rows", + label: "Score a list of rows.", + className: "api-method post", + }, + { + type: "doc", + id: "api/score-a-batch-of-rows", + label: "Score a batch of rows.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "ScoringFunctions", + link: { + type: "doc", + id: "api/scoring-functions", + }, + items: [ + { + type: "doc", + id: "api/get-a-scoring-function-by-its-id", + label: "Get a scoring function by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-scoring-function", + label: "Unregister a scoring function.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-scoring-functions", + label: "List all scoring functions.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-scoring-function", + label: "Register a scoring function.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Shields", + link: { + type: "doc", + id: "api/shields", + }, + items: [ + { + type: "doc", + id: "api/get-a-shield-by-its-identifier", + label: "Get a shield by its identifier.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-shield", + label: "Unregister a shield.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-shields", + label: "List all shields.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-shield", + label: "Register a shield.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "SyntheticDataGeneration (Coming Soon)", + link: { + type: "doc", + id: "api/synthetic-data-generation-coming-soon", + }, + items: [ + { + type: "doc", + id: "api/generate-synthetic-data-based-on-input-dialogs-and-apply-filtering", + label: "Generate synthetic data based on input dialogs and apply filtering.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "Telemetry", + link: { + type: "doc", + id: "api/telemetry", + }, + items: [ + { + type: "doc", + id: "api/get-a-span-by-its-id", + label: "Get a span by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-a-span-tree-by-its-id", + label: "Get a span tree by its ID.", + className: "api-method post", + }, + { + type: "doc", + id: "api/get-a-trace-by-its-id", + label: "Get a trace by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/log-an-event", + label: "Log an event.", + className: "api-method post", + }, + { + type: "doc", + id: "api/query-metrics", + label: "Query metrics.", + className: "api-method post", + }, + { + type: "doc", + id: "api/query-spans", + label: "Query spans.", + className: "api-method post", + }, + { + type: "doc", + id: "api/query-traces", + label: "Query traces.", + className: "api-method post", + }, + { + type: "doc", + id: "api/save-spans-to-a-dataset", + label: "Save spans to a dataset.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "ToolGroups", + link: { + type: "doc", + id: "api/tool-groups", + }, + items: [ + { + type: "doc", + id: "api/get-a-tool-by-its-name", + label: "Get a tool by its name.", + className: "api-method get", + }, + { + type: "doc", + id: "api/get-a-tool-group-by-its-id", + label: "Get a tool group by its ID.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-tool-group", + label: "Unregister a tool group.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-tool-groups-with-optional-provider", + label: "List tool groups with optional provider.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-tool-group", + label: "Register a tool group.", + className: "api-method post", + }, + { + type: "doc", + id: "api/list-tools-with-optional-tool-group", + label: "List tools with optional tool group.", + className: "api-method get", + }, + ], + }, + { + type: "category", + label: "ToolRuntime", + link: { + type: "doc", + id: "api/tool-runtime", + }, + items: [ + { + type: "doc", + id: "api/index-documents-so-they-can-be-used-by-the-rag-system", + label: "Index documents so they can be used by the RAG system.", + className: "api-method post", + }, + { + type: "doc", + id: "api/run-a-tool-with-the-given-arguments", + label: "Run a tool with the given arguments.", + className: "api-method post", + }, + { + type: "doc", + id: "api/list-all-tools-in-the-runtime", + label: "List all tools in the runtime.", + className: "api-method get", + }, + { + type: "doc", + id: "api/query-the-rag-system-for-context-typically-invoked-by-the-agent", + label: "Query the RAG system for context; typically invoked by the agent.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "VectorDBs", + link: { + type: "doc", + id: "api/vector-d-bs", + }, + items: [ + { + type: "doc", + id: "api/get-a-vector-database-by-its-identifier", + label: "Get a vector database by its identifier.", + className: "api-method get", + }, + { + type: "doc", + id: "api/unregister-a-vector-database", + label: "Unregister a vector database.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/list-all-vector-databases", + label: "List all vector databases.", + className: "api-method get", + }, + { + type: "doc", + id: "api/register-a-vector-database", + label: "Register a vector database.", + className: "api-method post", + }, + ], + }, + { + type: "category", + label: "VectorIO", + link: { + type: "doc", + id: "api/vector-io", + }, + items: [ + { + type: "doc", + id: "api/insert-chunks-into-a-vector-database", + label: "Insert chunks into a vector database.", + className: "api-method post", + }, + { + type: "doc", + id: "api/list-files-in-a-vector-store", + label: "List files in a vector store.", + className: "api-method get", + }, + { + type: "doc", + id: "api/attach-a-file-to-a-vector-store", + label: "Attach a file to a vector store.", + className: "api-method post", + }, + { + type: "doc", + id: "api/returns-a-list-of-vector-stores", + label: "Returns a list of vector stores.", + className: "api-method get", + }, + { + type: "doc", + id: "api/creates-a-vector-store", + label: "Creates a vector store.", + className: "api-method post", + }, + { + type: "doc", + id: "api/retrieves-a-vector-store", + label: "Retrieves a vector store.", + className: "api-method get", + }, + { + type: "doc", + id: "api/updates-a-vector-store", + label: "Updates a vector store.", + className: "api-method post", + }, + { + type: "doc", + id: "api/delete-a-vector-store", + label: "Delete a vector store.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/retrieves-a-vector-store-file", + label: "Retrieves a vector store file.", + className: "api-method get", + }, + { + type: "doc", + id: "api/updates-a-vector-store-file", + label: "Updates a vector store file.", + className: "api-method post", + }, + { + type: "doc", + id: "api/delete-a-vector-store-file", + label: "Delete a vector store file.", + className: "api-method delete", + }, + { + type: "doc", + id: "api/retrieves-the-contents-of-a-vector-store-file", + label: "Retrieves the contents of a vector store file.", + className: "api-method get", + }, + { + type: "doc", + id: "api/search-for-chunks-in-a-vector-store", + label: "Search for chunks in a vector store.", + className: "api-method post", + }, + { + type: "doc", + id: "api/query-chunks-from-a-vector-database", + label: "Query chunks from a vector database.", + className: "api-method post", + }, + ], + }, + ], +}; + +export default sidebar.apisidebar; diff --git a/versioned_docs/version-v0.2.23/api/synthetic-data-generation-coming-soon.tag.mdx b/versioned_docs/version-v0.2.23/api/synthetic-data-generation-coming-soon.tag.mdx new file mode 100644 index 0000000..0d1771b --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/synthetic-data-generation-coming-soon.tag.mdx @@ -0,0 +1,19 @@ +--- +id: synthetic-data-generation-coming-soon +title: "SyntheticDataGeneration (Coming Soon)" +description: "SyntheticDataGeneration (Coming Soon)" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/telemetry.tag.mdx b/versioned_docs/version-v0.2.23/api/telemetry.tag.mdx new file mode 100644 index 0000000..5ec6c2d --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/telemetry.tag.mdx @@ -0,0 +1,19 @@ +--- +id: telemetry +title: "Telemetry" +description: "Telemetry" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/tool-groups.tag.mdx b/versioned_docs/version-v0.2.23/api/tool-groups.tag.mdx new file mode 100644 index 0000000..3fcb9cc --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/tool-groups.tag.mdx @@ -0,0 +1,19 @@ +--- +id: tool-groups +title: "ToolGroups" +description: "ToolGroups" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/tool-runtime.tag.mdx b/versioned_docs/version-v0.2.23/api/tool-runtime.tag.mdx new file mode 100644 index 0000000..aa26332 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/tool-runtime.tag.mdx @@ -0,0 +1,19 @@ +--- +id: tool-runtime +title: "ToolRuntime" +description: "ToolRuntime" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-benchmark.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-benchmark.api.mdx new file mode 100644 index 0000000..f5cdfce --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-benchmark.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-benchmark +title: "Unregister a benchmark." +description: "Unregister a benchmark." +sidebar_label: "Unregister a benchmark." +hide_title: true +hide_table_of_contents: true +api: eJztV9FO20oQ/ZWVn1oJEopa9TZvtFBddKlAITxcUXQ1sSfxlrXXd3cdiCL/e8/YTuKQUOUVKbzYzs7szJw5Z3ZZRI59YXPPPhosotOTE3kk7GOni6BtHg2i63+i6ij6uGtplLJy/H/JPqgn8krnMzI6UdapjMzEuoyT6CiKbR44D+JPRWF0TOLf/+Vlk0Xk45QzkrcwLxjb2vEvjgMcC2cLdkE32flAofQdO41tp+xguJnW36PRjWqsVWwTlvyDDoY7vj44nU+3XC+cQ/K18ZEi5VPrgvJllpGbKztRARVzbfOU6jhVuinaacqDQsGK8na9jlPJ/oG02TdyYy2hjc1Rm0rLjPJjx5TQ2LDq+GykI5F0jprzeI8q313Xb2TeqzN1N7xqi4mR/JhV6TlRwaKz8OYZq8w6RpnSz7pzisa2DHVwX3CsJzpWNo5L5xjhX+SFxChJdBPvptPQCRnPR5HwRzvwZHC/7PCyWyvwHlb9a2B6Bbwll9XE2azO4ezmsqeGtpymZo72GGOfvBp+/6Y+/3XyuSeg8TNlRcOMJb9A9XW8r5SoYUPxdT57Mr+qpPqPp192Kyc2GqpQKdy9vARr4ZzPlxvLriDCVM8Yz8yWsBFsdcYHUR1E9cZEdfplHW8Epv8QprfK8l1p/WtLaAII8XPMnAhsIjYKrIzOdOipG8OEhALYQ1OCSAwWXa/V26fXTirPbobmA06REgMeIVaZ8zPwDvjiFoaDtA7SekPS+tQ9ry6F2chc3TZsX8dv5XW2TfklGElPXZdgAlNWn0pjxsGT2wDAsNbKK+EJlSZsS+xPGx9EdRDV2xLVya5oe0ioVolQjaZeqvkKNFIQ59FLAZsJ3+WOp9pDsGj1eGnYE86To4yxIHssohwfsF+Z/KdFUVr2KCikW0jIaXd5vuzByk36Wa5iSpw17sGV6MOWClvuVBWSR0KphSmCGeRWp4ngg6g/+0CmSKnPuAT3V+F8f9HNuIJDcwQ3RZVOsExDKAb9Pi4Dx6lFWsmxMZTRMfoQP/Zim0US2jPQ1WFeO54vJ9D9g6wJJ+uM23Zdib+6FX912zKzZiziS/AGotmHHaiJkP0Gpakrsu7OIaWgMJpmGpv8zNWLP0wNrm/rnCeFxYzyGAr1TUY7jCPXsDQBtkoGmJtQzL7ZlBz/zIVp1tUS3N58LP91GIjS0XSZ1Q+MOONrNhfAEfNCMGmJ8zrRNhBYrIf0H1xaZgR+Dv3C4PolMetmLlpC3EctIWAtlMBjTQp8DDaIjBZK58VtsRjjYnfnTFXJz7gbOmk5XuvpOpb+ggCJ9vKerMbHq1W8G7YEf6/21sTO+tofQVNhEZlSvvD6yPOXwqweYJ9iVuNQkmwbk7M45iJ0nLcOwQ2JnV9cXYwuoLvfciElXQ== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a benchmark. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-dataset-by-its-id.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-dataset-by-its-id.api.mdx new file mode 100644 index 0000000..d390191 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-dataset-by-its-id.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-dataset-by-its-id +title: "Unregister a dataset by its ID." +description: "Unregister a dataset by its ID." +sidebar_label: "Unregister a dataset by its ID." +hide_title: true +hide_table_of_contents: true +api: eJztV9tu20YQ/ZUBnxJAlhwjQRq9uZWLGnVgQ5YfCscoVuRI3GTJZXeXsgVB/94zvFiUJSd6NSC/iJe5njlnll5Fjn1hc88+Gq6is9NT+UnYx04XQds8GkbXf0frXvRx36tJyuT4v5J9oEflSecLZXRC1lGmzMy6jJOoF8U2D5wH8VdFYXSsxH/w3UuQVeTjlDMlV2FZMMLa6XeOAxwLZwt2QdfV+aBC6Tt2GmHn7GC4XdZfk8kN1dYU24Sl/qCD4Y6vD07n8x3XC+dQfGXcI0U+tS6QL7NMuSXZGQV0zJXNY6rjlHTdtNMqD4SGSeXN+yrPWuIHpc2hmWtrSW1sjt4oLTOVnzhWiZoapo7PVjmSSefoOY8P6PLddXWlzHs6p7vxVdNMjOKnTKXnhILFZOHNC6bMOkabMs9qcqSmtgxVcl9wrGc6JhvHpXOM9C/qQmEqSXSd76Yz0JkynnuR8Ec78GR43064ndYzeA/P86thegW8lss0czaraji/uezT2Jbz1CwxHmPso6fxn3/Q599OP/cFNH5SWVEzo+UXqL7J97tKaFxTfFPPgcxfr6X7j2df9isnNhqqoBTuXi6CtXDOl21giQoizPWC8ZvZEjaCrc74KKqjqN6YqM6+bPJNwPSvwvRGWb4rrX9sCU0AIX6KmROBTcSmApPRmQ59ujGsUFAAe9RcQSQGL12/0dun104qz26B4QNOkRIDHiFWmfMT8A644waGo7SO0npD0vrUPa8uhdmonG5rtm/yN/I636V8C0bSp+sSTGCVVafSlHHw5DYAMLxr5JXwTJUm7ErsZ4GPojqK6m2J6nRftgMkVKlEqKbmXroZqYCzCifcw8ty73LHc+0hVww6qc1ouiSND7/LUV+Yr5zKGAYSaRXluIFfY/qvFlVpiVSokO6gISfe5aidQxsfEy2f80qODfLBlZjEjg4b9qzXaADFpDaRGtigrqpEpB5Gg8WHQZPCD1abCtcwqY/duoXSCX5pCMVwMMAHwElqUUhyYozK1Amwj3/0Y5tFkswzENVhWTmO2q1z/yDvhIdVjc2IrsSfbsWfbhs2VixFfkleQ7L4sAclEa/forHqCqsbOaQqENbRQiPIt5xe/GFTcPWFznlSWOwlj0VQfb1ohxXkamYmQJNkabmZitnXQZXjb7mwy7pKdrvBp/KfhoEQnZq3VX3FWjO+YnABHLEjBJOGJr+m1xYSq82CPsC14UbgpzAoDD7BpIZquKuGEvcN2i39e9GwQ1yMUCYvZqvVFE/vnFmv5TG+B52MHJfVRp3KfEGARHu5Tp5XxqvVvxs3lH5PB2pgbzfNQ5BUOKRMKXe4/MHLbRGuH2CdYjfjEJJKa4PzOOYidFx3Dr0tQY0uri4mF1DZ/xImHW4= +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a dataset by its ID. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-model.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-model.api.mdx new file mode 100644 index 0000000..a882745 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-model.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-model +title: "Unregister a model." +description: "Unregister a model." +sidebar_label: "Unregister a model." +hide_title: true +hide_table_of_contents: true +api: eJztV9Fu2zYU/RVCTy2Q2GnQoqvfsjXDiqVI4DgPQxoMNHVtsaVIjaScGIb/fedSki3XzpA9BnBeTIm8vOeeew6prDJPoXI2UMhGq+z87Ix/cgrK6ypqZ7NRdv1ntj7J3h+amhQkPP1TU4jiUQah7UIanQvnRSnNzPmS8uwkU85GspHjZVUZrSTHD78H3mSVBVVQKXkUlxVhWzf9TioisPKuIh91gy5EGevQW6ex7Zw8Fu7C+mMyuRHNaqFcTow/6mioFxui13a+F3rpPcCnxSdCilA4H0Woy1L6pXAzEVExpTWPhVaF0E3RXksbBQoW0rbzKc+a949Sm5dmblZzauMsahNFXUp76knmcmpI9GJ24HAmbVGzVS+o8s11GknzVlyIu/FVW4wC+CmJOlAuokNnEU0LEqXzhDK5n6lzQk5dHVPyUJHSM62EU6r2npD+J1wAJvNcN/lueg2dSRPoJGP9aA+djO67Dnfd2pD3sOlfQ9Mz5HVaFjPvyoTh4ubLQIxdPS/MEu0xxj0GMf79N/Hxl7OPAyaNnmRZNcro9AWpb/P9KnMxbiS+xfNC5a/XXP3780+HnaOMhitEgfDAg+gcgu2y25h3hRDmekH4LV2NNcytLuloqqOpXpmpzj9t802g9K+s9NZZoW+tv1wNT4AhelJEOdPGZpORhNGljgNxY0gCUIR65FzCJAaTftD67cNzN1Ugv0DzQSdbiUAPC6u29AS+I56opeForaO1XpG1PvTvqy+sbCAXt43at/lbe13sS74jIx+I6xpKIFmmW2lKuHisiyAMc629cprJ2sR9i/3XxkdTHU31ukx1dijbCyyUXMJSk/PA1XyF9Exg8Ltg76ynuQ4wK9pc8qIBa116WRJecuwqs3jA2jT9t2YXaY6tZCz2qucbTudwGLvVd/ynUO5jvcnHebZ8R1+D/z33tZpZrwEcgAqHpUhogC3BBIBRNly8G6YEYbjqMK4x3Vy0TQm1Z8aKGKvRcIgr/7RwAJGfGiNLeQq21Y+BcmXGiQKBQx2XKfBzd87cP/AcKy/ha5tyxfHiluPFbau/pEvk5+QNKYt3B3hiu4Yd4cq+lfo7x0JGgQNoAWbDNyt++sPZQOmbnGxeOZxEAdZP3yva49DxjRZzMCn4mPIzqSg0m0pP3yzryflktP3Np/y/hYH1vJx3qBo1Jc1W4BGnAnPSyuSwpHaqX22P4WeWt/2P9BSHlcHHFedKTVy1bb9vWG0aj8FoI0+0ibvLS1arKT7R7rxZr/k1vvI8txXDdE5OuYdocq4Dj/PNQfAs2jfjVrJvxf9S+sF62peQI6tFmpqfMPxBy77d1g9YW+DUxfXCaJvpC6Woir3AvetsxzSfL68uJ5dw0r/xABFW +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a model. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-scoring-function.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-scoring-function.api.mdx new file mode 100644 index 0000000..b1a5432 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-scoring-function.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-scoring-function +title: "Unregister a scoring function." +description: "Unregister a scoring function." +sidebar_label: "Unregister a scoring function." +hide_title: true +hide_table_of_contents: true +api: eJztV11PGzkU/StX89RKkFDUqtu8sYVq0VKBIDysKKocz03GrceetT2BKMp/33NnJiQhsKWPSOElM+P7ce6959hmngWOlXeRYzaYZ4cHB/KTc9TBVMl4lw2y87+zxV72/qmlYcEU+N+aY6I7Fcm4qbImJx+oVHbsQ8l5tpdp7xK7JP6qqqzRSvz7P6IEmWdRF1wqeUqzihHWj36wTnCsgq84JNOii0mlOq7ZGYSdcIDhJqy/hsMLaq1J+5wFfzLJ8ppvTMG4yZbrSQgA3xjvkaJY+JAo1mWpwoz8mBIq5sbmrjC6INMWHYxyiVAwKdetN3kWEj8pY1+aubWW1NY71EZFXSq3H1jlamSZ1nw24Egm41Cz0y+o8s1586TsWzqi68uzrhgN8COmOnJOyWOy8OYpU+kDo0yZZzM5UiNfpyZ5rFibsdHkta5DYKR/hAvAVJ6bNt/F2kDHykbey4Q/JoAng5vlhJfTemje7cP82jY907wll2kcfNlgOLo47dGlryeFnWE81vq7SJdfPtPHPw4+9qRpfK/KqmXGkl+g+irfnyqny5biKzwvZP5iIdW/P/z0tHK0NVAFFXCP8pC8h7ObLQNLVBBhYqaM39LXsJHempJ3otqJ6pWJ6vDTKt8QTP8qTO+UFdel9Y+voQl0iO81cy5tE7GpxGRNaVKPLiwrAEpgj5ooiMRiMfQ6vX147qSKHKYYPtopUmK0R4hVO75HvxPeuGvDTlo7ab0iaX1YP69OhdlATlct21f5O3kdbVN+2Yy8R+c1mMCqbE6lEePgcT6hYVjr5JXzWNU2bUvs/wLvRLUT1esS1cFT2V4goUYlQjU1iVLNlfbS/S+10wIzShmbsK9d4ImJkK1wrTWncWffEwGooErGugScZw4vMtXW8vvYfTeiLyOxKpWKrb7I2Xd6vJzI4wwy5PoBguRbDSOFGsPZkmZHqMUCtQBY4WGKnBYYG7jAMMj603f9Ltf+MlfszzdgL2DensptaXWQ9hYpVYN+H/eD/cIDVL5vrSrVPkajf/a0LzNJHBkNN2nWOB4vN6WbW1kTmjZ4uwmeiT9diT9ddWRtSIz8krzt0/TdE60TbccNlqt13a1HToVKhN1qahDkm6NHfxguNxd4dnnlsW1F7BPN5cYE7FChJW4us5E9LYyV5tgGVYG/OSGfD40qt4OP5B8RC50GNVmi+opdz8aG4BX6iC1EetLR55es22jEfLV9/9qzY0ni+9SvLO5ngqAZ7bwjx03b6y164Ntgk9cYpnBAXObzEW5918EuFvIZF8cgw8djs/WOZNKgQm6iPOcPe8uzhby57Ij+ln5XIk+W2H0Eb4VWytbyhsefPNvS6+IWDgU2dJxcgrq1OdKaq7TmvXVSbkju+OTsZHgCHf4HJbc1sA== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a scoring function. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-shield.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-shield.api.mdx new file mode 100644 index 0000000..c066a7b --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-shield.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-shield +title: "Unregister a shield." +description: "Unregister a shield." +sidebar_label: "Unregister a shield." +hide_title: true +hide_table_of_contents: true +api: eJztV8Fu2zgQ/RVCpxZI7DRo0a1v2W0WW2yKBI5zWKQ50NTYYkuRWpJyYhj6976h5Fiu7UX2GMC5mLZmOG/evEcqq8xTqJwNFLLRKjs/O+OPnILyuora2WyUXf+dNSfZ+32PJgUJT//WFKJ4lEFou5BG58J5UUozc76kPDvJlLORbOR8WVVGK8n5w++BN1llQRVUSl7FZUXY1k2/k4pIrLyryEfdogtRxjr04jS2nZNH4DasvyaTG9FGC+VyYvxRR0O93BC9tvOd1EvvAT4FnwgpQuF8FKEuS+mXws1ERMeUYh4LrQqh26a9ljYKNCyk7Z6nOg3vH6U2L63cRnNp4yx6E0VdSnvqSeZyakj0crbgcCVt0bNVL+jyzXVaSfNWXIi78VXXjAL4KYk6UC6iw2SRTQsSpfOENnmeaXJCTl0dU/FQkdIzrYRTqvaeUP4XXAAm81y39W56A51JE+gkY/1oD52M7tcTXk/rmbyH5/m1NB0gb61lMfOuTBgubr4MxNjV88IsMR5j3GMQ4z//EB9/O/s4YNLoSZZVq4y1viD1Tb3fZS7GrcQ3eF6o/Kbh7t+ff9rvHGU0XCEKpAdeROeQbJfrjXlXCGGuF4TP0tWIYW51SUdTHU31ykx1/mlTbwKlf2Wld84KfWv942p4AgzRkyLKmTY2m4wkjC51HIgbQxKAItQj5xImMXjoB53fPhy6qQL5BYYPOtlKBHpYWLWlJ/Ad8Y06Go7WOlrrFVnrQ/+++sLKBnJx26p9U7+z18Wu5Ndk5ANxXUMJJMt0K00JF491EYThWWevnGayNnHXYv+18dFUR1O9LlOd7av2Agsll7DU5DxwN7eFJpMHRr+N9s56musAtyaJcdSA1S69LAm/cvYqs/jCIs9hHbYh49CcXslY7DDAt9wmdD2DdnMeZv1ck0ttSI++xhB2LNgJp2kAHpgKh1BUNICXkALBKBsu3g3bCmG42hRvENFeuG0jtWfmihir0XCIq/+0cMCRnxojS3kK1tWPgXJlxrUCgUsdlynx8/q8uX/gZ6zABLEbzhXni1vOF7edDpM+UZ+Lt8Qs3u3him0btgQs+5bq7xwLGQUOogUaDN+s+OUPA6T0bk42rxxOpIAjIL23aI/Dx7eazEGm4OPKz6Si0G4qPX2zrCvnk+F2N5/y/xgGFvRyvkb1FQeaCUm7FXjE6cCcdGI5oKyt9leb8/hQfCeCSE9xWBm8ZnG1NMZVN/v7ltdu+liNejrFrHjEHLVaTfG+dudN0/DPeOXzPFss06E55UFi0rkOvM6fT4WDiN+MO+m+Ff9P8nt76n6EKFkz0tT8DcsftNy2XvOA6AKnMNYMuA24UIqq2Evdud62/PP58upycglT/QRnKRgu +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a shield. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-tool-group.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-tool-group.api.mdx new file mode 100644 index 0000000..f19bd94 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-tool-group.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-tool-group +title: "Unregister a tool group." +description: "Unregister a tool group." +sidebar_label: "Unregister a tool group." +hide_title: true +hide_table_of_contents: true +api: eJztV11P20gU/StXfmolSChq1W3e2MLuoqUChfCwoqia2DfxtGOPd2YciCL/955rO4lDAsorUniJE9/Pc8+5Mywix76wuWcfDRbR6cmJfCTsY6eLoG0eDaLrf6PqKPq469UoZXL8f8k+0KPypPOZMjoh6yhTZmJdxkl0FMU2D5wH8VdFYXSsxL//00uQReTjlDMlT2FeMMLa8U+OAxwLZwt2QTfV+aBC6Tt2GmGn7GC4WdY/o9ENNdYU24Sl/qCD4Y6vD07n0y3XC+dQfG18RIp8al0gX2aZcnOyEwromGubx1THKemmaadVHggNk8rb93WeSuIHpc2+mRtrSW1sjt4oLTOVHztWiRobpo7PRjmSSefoOY/36PLddf2kzHs6o7vhVdtMjOLHTKXnhILFZOHNM6bMOkabMs96cqTGtgx1cl9wrCc6JhvHpXOM9M/qQmEqSXST76Yz0Ikyno8i4Y924Mngfjnh5bRW4D2s5tfA9AJ4Sy7TxNmsruHs5rJHQ1tOUzPHeIyxj56Gf32lz3+cfO4JaPyksqJhxpJfoPo6358qoWFD8XU9ezK/qqT7j6dfdisnNhqqoBTuXh6CtXDO58vAEhVEmOoZ4zOzJWwEW53xQVQHUb0xUZ1+WecbgenfhOmtsnxXWv/ZEpoAQvwUMycCm4hNBSajMx16dGNYoaAA9qipgkgMXrpeq7dPL51Unt0MwwecIiUGPEKsMucn4B3wjVsYDtI6SOsNSetT97y6FGajcrpt2L7O38rrbJvySzCSHl2XYAKrrD6VxoyDJ7cBgOFdK6+EJ6o0YVtirwU+iOogqrclqpNd2faQUK0SoZqaeukGR53529my8NLAZsF3ueOp9hAsRo3bn6GpWPaE9MqpjPFGgiyiHF/gIDa1yQ8tktISpFAh3YJCjrvL8+UQ1qFlouUqqyRaIx9ciUls6bBlT1WhfFSUWpgim0FxdZ3IPoj6sw/9VXG+v+gWWsGsOXqbXkonGKYhFIN+H5eA49SimOTYGJWpY+Af/+rFNoskoWegqsO8djxfbp77B3knXKzrbMd0Jf50K/502zKyZiryS/IGmdmHHWCJgP0GlVVXXN3IIVWBsJJmGkG+5/TsD9uC61s650lhsZs8lkF9g9EOa8g17EyAKMnichMVs2+CKsffc2GYdbX0toOP5b8NAzE6NV1W9Q2rzfiaxQVwxJ4QTFq+vEKwDQgW6+38mk/LiMBPoV8YXLwkaz3ORUuE+wbfNRXwZbDBWgxO5i2mi8UY17g7Z6pKfsZN0Mmg8Vjv0rFMFWNPtJfnZLUsXiz93bAl83vaXwA7m2p/BDuFPMqU8g2Pv3j+XIbVA+xTrGacQVJuY3IWx1yEjvPWmbehp/OLq4vRBUT2G8MVIOI= +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a tool group. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/unregister-a-vector-database.api.mdx b/versioned_docs/version-v0.2.23/api/unregister-a-vector-database.api.mdx new file mode 100644 index 0000000..9ec16a0 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/unregister-a-vector-database.api.mdx @@ -0,0 +1,68 @@ +--- +id: unregister-a-vector-database +title: "Unregister a vector database." +description: "Unregister a vector database." +sidebar_label: "Unregister a vector database." +hide_title: true +hide_table_of_contents: true +api: eJztV99v2kgQ/ldWfmqlAGnUqlfe0ianqy5VooScdEqjavAOeNu117e7JkGI/73f2CaYQNro3iLBCzaeH9/MfN+sWSSeQ+mKwCEZLpKjw0P50hxSb8poXJEMk/O/k+VB8nbXo1HGyvN/FYeo7igoU8zIGq2cVznZifM56+QgSV0RuYjiT2VpTUriP/geJMgiCWnGOclVnJeMsG78ndMIx9K7kn00DboQKVahY2cQdsoehpuw/hqNLlRjrVKnWfBHEy13fEP0pphuuZ56D/C18YEiFTLnowpVnpOfKzdRERVzbXOXmTRTpinaGyqiQsGKivZ5nWcp8SMZ+9zMjbWktq5AbSqrcip6nknT2LLq+GzAkUymQM1F+owqX53XV2Rfq2N1fXnWFpMC/JhVFVir6DBZePOMVe48o0yZZz05RWNXxTp5KDk1E5Mql6aV94z0j3ABGGltmnwXnYFOyAY+SIQ/xoMnw5vVhFfTemje7cP8mjY90bwVl9XEu7zGcHzxua8uXTXN7BzjsdbdBXX55yf1/o/D931pGt9TXjbMWPELVF/n+0haXTYUX+N5JvOXS6n+7dGH3cpJrYEqVAb3IBfROTgX81VgiQoiTM2M8Z27CjbSW5PzXlR7Ub0wUR19WOcbgelfhOmtskJXWv+6CppAh/g+ZdbSNhEbRVbW5Cb21YVlAqAI9tCUIBKLh77f6u3dUydVYD/D8NFOkRKjPUKsquB79Dvijts27KW1l9YLkta77nn1WZgN5OqqYfs6fyuv423Kr5qh++q8AhOY8vpUGjMOnsJFNAzPWnlpnlBl47bEfhV4L6q9qF6WqA53ZXuGhGqVCNVoGqSaf2Dk/MnHIPg38V4XnqcmQK+Y9Ky2U5oijXG49YX45ClnPJZAi6TADbwaw296/M2IrIxEKilmW+2QI89oSE7k61cDeZRHRlw94JCs61FEX2E0W8Js6bRcoiDAyxxMkdoCaQ0aUIbJYPZm0KTq6XEYLLqolzBrzuKmsMpLU7MYy+FggLeCXuYARvespZx6GEj6o5+6PJGEgdFmE+e148lqFd3cyjMhZ42znduZ+Ksr8VdXLUVr6iK/JG/aNHuzo3Oi6LDBbeqqrRs5ZhQVdtQMvQ5fC/Xog/XB9Ws7F7p0WFYB26F+pTEee8k3dNXoqJJN5ieUcmiCkuevhVDO+VqL28HH8vfDQp2epitUX7DrbKhpXaKPWBzSk5Y8v6PcRh8W6539W8eWG5Hv46C0eCeT/PVgFy0lbppOr0mBm+EGmTFCmbyYLhYS+Nrb5VJ+xkuil5Hjsl6zY5kvCKBNkGv9sEeexP/qsqX1a/U/dbGzwvZHkFZKI1vJHS5/8PyxVJe3sM+wwnFWCfbG5DhNuYwd562zcUNmJ6dnp6NTaO8ntbAvFw== +sidebar_class_name: "delete api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Unregister a vector database. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/update-an-existing-prompt-increments-version.api.mdx b/versioned_docs/version-v0.2.23/api/update-an-existing-prompt-increments-version.api.mdx new file mode 100644 index 0000000..1bfd691 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/update-an-existing-prompt-increments-version.api.mdx @@ -0,0 +1,68 @@ +--- +id: update-an-existing-prompt-increments-version +title: "Update an existing prompt (increments version)." +description: "Update an existing prompt (increments version)." +sidebar_label: "Update an existing prompt (increments version)." +hide_title: true +hide_table_of_contents: true +api: eJztWG1r20gQ/iuLvjSB2E5LS9twd5D2elyhJSFxCkdqyloaW9tKu+ruyo4x/u/3zGply3FyTUrhCCQfEkUazevzzGhnmVhyldGOXHK0TJ4dHvKfjFxqVeWV0clRMsxJ1FUmPWXi1Jqy8gIvmdqmJObK50Lp1FJJmgVmZB1e6ycHSWpwR3tWKKuqUKlkhYOvjrUuE5fmVEq+8ouKYMeMv1Lq8WJlTUXWq8anKpjsyDlvlZ5CbtdNt3CeStG8Ijxd+cbDmbRKjgsSVSFTyk2Rwc2++BRvOyEtCaOLhXB1VRnLkcxz0qJ2MCU8VJ+1eRLHp+/7yeogiaF2PFMIeEp2x7VPjaTYiwLCeYn4oFl68fRgK4EQc3JG+2yhieOLyn4c/YVW32sSKoMWNVGwMTG2lJ5VSieeVKzot+evepmaKt/Lpcv/eBKiaHPQsSGtlQuYUEim27W9um78g3JemEmb93W2tSyRL58jylRqMQaOHPxROmR0XaUSVfHEapX7ktFE1kW33mNjCpI6hByfTWTh6LoXbxo5qM8C1pBe1BCWLMwp10JTKBfMR2Xr28hXIxcBt4I/MssUa5fFaQeT0bql77WyhOJcrrHQrVk3t1uhjQ4Sr3zBsTV82qnmcZucNdEsVbjm4jJqACADy+KkIn38XryFLAIOAN9OKuf6QyFLKc69TL8BtyuO6/ltPOeYCMWcAzNKz2ShgEgrSlkwnCj7dbQGBXzt7kCev4fDU9FIi9RkAScxfT8ixTtruagsfMBJy8FsMLwspV0wXhkGFGTmuUpzRgYHjaJpH/AANDXPg52Aey9VcVfLjTSbLoxm3ud1KXXPkswCPzrvbLkTqKARs07vEOXeSdVgdF8ci4uzDzGYLuO8QWXxNs1IlEAOwmzaA1uWY1P7YNxVlKJ5pMKkaW0twfw1v+7DiVjhtlrr5HXgH9J0S/LawSQmgHTwgRuvODP1NEejnpiiMHMnzv56K16+OnwZWjJdSeCeuvgC1Df23sgMjTxAfOPPHZEfmfPs9c3MSQsFVgg0VsE0RcoNXtaLVjFrBRCmaoaxIktT69AyvSrpkVSPpHpgpHr2emNvCKR/ZKRHZrkutf4xNTiBDNFVSpRx2phsPJoKVSrfF6cY2nDIAz1yKkESnlu2nVQvbptUjiyGrkA6mUrE0xC1qTVdId/81UMxDY/UeqTWA6LWi+68es/IhufivEH7xn6k1/Eu5NtkZH1xUgMJJMswlcaEwaNNOB1kLb06X9vXvkD/Q/EjqR5J9bBIdXiTtTtQKLCEoSanjqNpjmuOvb92/A77kQCVKxyG+ZAWT2J768P9+gS6z9uRSlqcjkFvVrxM+KgMPd3To2LFONflN646Oif9WJ326GfitoatbErhbY3S7BCzPdKvRo0w5vcbky1Y4n/Y37Rrpu4CJ7pxz5VL+CQP8N2c8LfzNCauUrTY/1WrkIsYQRFXIj+3CwnuOPJf5I8WIl3b59QwWNN8HbPcXnbsxYvfh4DDfv9+XK7aZcVm23HNxw6rm0Q0hGmPXIFL25DEDbAgNxlj3YRjWYD8UTKYPR00Ft1guebFKhjlYdjwprbM4dz76mgwwEdoL4cSynoFLz16Liw9UlMmjG9HQITyi/Din21aL0f8jFF9tsH/u00PafG7we0ahYdbmLlsJUa7tYuhcscNhYxJ6qxmxHnsu4FlXSvJ7OkN8OYx5bYatuyOkK7mADqEMUPPcJ+1uPaDmUgBq6SzyihuVFKH73RlwT7b9OCM2cJssxOZtkiWlj5r7qNhG+XNrvIxn6kLjBwrp61XHzHACxcwziXHNOScxA54/066lZlO2/oJVZFg3HcGICI6MHwMEFtGWF421YjAxNXRpmWj7ow+FlouxzjUXNhiteLbwJVl2I02gAkgzJTj6+zmjWY3lr2zSJt9cffuf2M4bWPTi7ChLGr+D5ffaLE1gFYjCOf4SkF3ZV+b528bj3pD1rJ5f2dArA7aN47TlJqucavsqNMETk/OhxAexyFU8sfeUWLlnKcZfgdfTchQ6Mfh3jIppJ7WgBieNzr551+SxLzX +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Update an existing prompt (increments version). + + + + + + + + + + + + +'"},"variables":{"type":"array","items":{"type":"string"},"description":"List of prompt variable names that can be used in the prompt template"},"is_default":{"type":"boolean","default":false,"description":"Boolean indicating whether this version is the default version for this prompt"}},"additionalProperties":false,"required":["version","prompt_id","variables","is_default"],"title":"Prompt","description":"A prompt resource representing a stored OpenAI Compatible prompt template in Llama Stack."}}}},"400":{"description":"The request was invalid or malformed","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":400,"title":"Bad Request","detail":"The request was invalid or malformed"}}}},"429":{"description":"The client has sent too many requests in a given amount of time","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":429,"title":"Too Many Requests","detail":"You have exceeded the rate limit. Please try again later."}}}},"500":{"description":"The server encountered an unexpected error","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":500,"title":"Internal Server Error","detail":"An unexpected error occurred. Our team has been notified."}}}},"default":{"description":"An unexpected error occurred","content":{"application/json":{"schema":{"type":"object","properties":{"status":{"type":"integer","description":"HTTP status code"},"title":{"type":"string","description":"Error title, a short summary of the error which is invariant for an error type"},"detail":{"type":"string","description":"Error detail, a longer human-readable description of the error"},"instance":{"type":"string","description":"(Optional) A URL which can be used to retrieve more information about the specific occurrence of the error"}},"additionalProperties":false,"required":["status","title","detail"],"title":"Error","description":"Error response from the API. Roughly follows RFC 7807."},"example":{"status":0,"title":"Error","detail":"An unexpected error occurred"}}}}}} +> + + diff --git a/versioned_docs/version-v0.2.23/api/updates-a-vector-store-file.api.mdx b/versioned_docs/version-v0.2.23/api/updates-a-vector-store-file.api.mdx new file mode 100644 index 0000000..1ec5d57 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/updates-a-vector-store-file.api.mdx @@ -0,0 +1,68 @@ +--- +id: updates-a-vector-store-file +title: "Updates a vector store file." +description: "Updates a vector store file." +sidebar_label: "Updates a vector store file." +hide_title: true +hide_table_of_contents: true +api: eJztWm1P20gQ/isrf7pKQELVXlu+pbRV0bUFQTjpBCja2Jt4y3rX3V0DOZT/fjOzdmLHDoQrX5CCREns3Xl9npl96X1khcuNdsJFB/fR634f/yTCxVbmXhodHUQD9reIvbFn8Cu+SCWOxz/hAbMih8lCe6mnzKeCFXnCvUjYBMbsRTtRbLSH1yiR57mSMUeJvZ8Oxd5HLk5FxvGTn+UCFBmSCxNza3JhvQxGyaQ2xnkL6mBM08ZzLX8VgskEzZlIYdnEWDIKjYnmO5XwTkkTXih4Fd2QnyOHju7RxFU9pesooqZsh3F1y2eOXbZFXEaonXvQNi588GjVX54kEhVwddLw3GhxPIkOLhYzdKEUiFt8HxujBNf1R7rIxsLWn5SO1p5wa/ms/qC0ZH41X3X4LzHbveEKYrt0gXHnTCwp17fSp404x2mhr0EfhMDCiOmsDaiz8g0rHKIFEuUAHX6BIxTFpPaGkTAHEWpFYg1Wwtt2jgGKDjPMC28aKV886DQQxdSSi4MvI7IYP2aA55hVDkfz+bpMTrhyYgeo9quQVgCcL4KBVzuRl16hpTWKHZYCKysGXSYOWvpZFXCyL+CQEQ4poG6vI+FPDZ/zqLIRwNqjTUIYhpdBDF9qEdypxLUwc2j0RE4LSyWE5dzyTHhh3YLnK7IW0QDLHvaZZozMjbCK5yNvroWukxRwKKZAqJrPb/r9VXd/EOuYmbAgAP6wUiQbC38rhGY8+cljKBkVqsHbjN+Ngn4n/xUbKX/fVv6d38msyJheNQK8DMp2WFY4D5YsjNnv9xnXCXvT//Dn05DbGa91rmyG8DPKXUjx/6DRAjVP0daB2DV0ohIXV/gbQ21awg9IBUoTiYIyqTnow/SVCJv9gGEgurQygy6IMMd+iJx+tPNtq9mD1axWLLZlbVvWtmXtOcsa2ruBVlryWYFrwRH3a3Jct2coMwF+ZTm7TSFhi/XeLYd1ZZLAghAwhk/rnEctijs/EtaGCtsU+sdxHuL6in3GEbB6BAJlgVVyEjQAP2LhHAZhwuFB8jiHTCI61+HtCiIsIKc0r732Xg7EqI0U9Arw5S4WAhwOLaThT3AC1YMnCW2cyrUx7TyACuhBEQID8XR82lngmlK/FhnXu5CshNJNxrJyNgtDx4s1eCX/SRzCeC0N6sYt7iC/QS4/l7EilBeuozYW1mJdqeUtDKVSUG05OvYGraDHJssVADt5MDVSj0DVFPa07sFxMdexUOoRaSXE5uuDcBb8hggUGK7ReNbcHnZUyVaNRGGYPZdDDQ77qTE0p1S6ahfFgljQUt+ajjbZVB99qiLd6L+4redSB6SUip6GE4nUW+5+l3vjru1jo74s0NIMWtu59WE/rvSu7OxzoQdH5VEHo9EMh7NgJ1VE8PFN1wnJEEKE/kFhozomNWyZZcKQXlxhJaJi80znIUu+PFJqvw6HJxVliJrLkv5Y6kMBosGwymEuNdYzV2QZt7MKFKF+3KYyTpkMTlvJgbC0otTle9JDxwoAGrWp5jAaVSujwTeWNmtXbU7DHNQkgYBI0cd11drGgJ2ffiudAYLjmoLYBM3ICpgtbgTLEBP1zsLHpvBhmZaLWE6gz5o4prIVixW7nsKPBcZDthbBq4E6VM/u4FVHemxiTUY2DE6O9tipKaapwgW1UubWsdMvh+zd+/47WkqLO441so4vWg9W+j7yhJ0GiC/t2RD5JXNef+hmTqwklvkUpuNpIoTcwGQ9qwSjVADCVN7gajMzBYzB2MJCYkuqLaleGKlef1jqGwLSvyPSS2a5OrX+MQVwAiJUrRNJKTZFRuvHPXaiBAeDPKCHT6EjMwUvbdWp3q7rVGGtyiCcSCUB4UFgFVrcQbzxTFeUYdhSa0utF0Stt/V+dYTIBsvZWUD7Un9Jr0Eb8lUwkj12XAASBM+oK43xZEEbuutJKnot1uSt67IHBG9JtSXVyyJVv0vbBhQiliDU+NShN2FfdXQctc46zunW2EGWW6e8eIm8PJSiLb4OJ+qr2z1MODzOuU/bJ07g+yZb2fIgCpIdLrJR+zIh3haQoBY9q8tVPAoobUMpT7Ppd/RehcGwevhoklk4tXqWCvNir6yHtf+KcN11fQ2RDulvXF7vPY2etfDUGImnCFwGSK+cO1S7J6JFM790iOdTkyBcDO2wCDUHUe9mv2dIJn6qo9717ldIMO/RzUjvvgTgHE9LqPsF6hQWSZt6nx/0erDq3E1BlUh2leIZ3wXCx9d7scnoyMoJoLH0M5r4qWp1F1f4DoF0uoTc52XRaEBmTlV4Eq65yvB8Q1XsDFWxs7JaEjrpBMe6kMCb/Q6+YHNxjTLL64W/LtmnnE4NbyQIudRs5Qc6maAdpNBJbqBvOroyACkSz1xtqJwJFgVsqnbCYwQNCuVWXGLFUMZSW2gLH+NOWEGjoGO5YNV3aLsq3FhhdqGHYUzKavFI+WuEocbtx+aVLPHizvdyBTWOzhkt9d6ArYsQ54AujP/+yjEa9p6DdqElkOGrqs4BJBBJKPH+fgw7knOr5nN8DBixCCH4SMuCMYIAAJVIh5+TBbHWevnHaUmUV+x3ynhnNKrKovGMkYoEfIOPUDM6OgyWnmew+1lsqyJPxS+F5RBUT4xreHsYrNgdhkukanarF6BHYcYgjkXuHxx7VStRJ8dnQxg8LvtNRrckkeW32LjgX7LUUFSoJ9Cz+0hxPS3ooiIKMvHnP5IymLg= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Updates a vector store file. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/updates-a-vector-store.api.mdx b/versioned_docs/version-v0.2.23/api/updates-a-vector-store.api.mdx new file mode 100644 index 0000000..723deb1 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/updates-a-vector-store.api.mdx @@ -0,0 +1,68 @@ +--- +id: updates-a-vector-store +title: "Updates a vector store." +description: "Updates a vector store." +sidebar_label: "Updates a vector store." +hide_title: true +hide_table_of_contents: true +api: eJztWV1PGzkU/SvWPLUSBKi22y1vlLZapLYgoCutCoqcGSdx67Gntic0ivLf91x7JnGSCaRqHxaJSJD5sO+Xz7nXvpllVrjKaCdcdjzLXhwe0lchXG5l5aXR2XF2wv4RuTf2Cn/ifPAVN8yKChOF9lKPmB8LVlcF96JgkzCUORrby/ay3GiPYSSVV5WSOSepB18diZ5lLh+LktOVn1YCykyQj4mVNZWwXkbDZJGMcd5CLcas2vlZy++1YLIgs4ZSWDaEJWRcalQ232uVdEoc8lrhVRbn9OOcdVVNFGh2om+PcXXHp47drMy+yUhnbgUFqM9TvRLBGQm7If9alsJ5Xlbsbiz0hg/sjjvWCCTZmpfi4fg8Ow9XXD1nnzCBmWFncGrHR6I/mPoY+Q5LmxAdrisggGAucxXPgQgHOAymm8ZLzaJ0KBtKJfq5qbV3m8B7j5cMOMiFc4QzRMTXcDwM717cvQdwlJuyUoLC9vAifKrLAUCEMJGVDsq4Z2M+EWwgsCiuzsmwYa3UtLUyLkfOdS6U+hUlBbkmber9UirFjctfkB9nM29a+SRS6j7uRpZuf15uXlsLHiAUA0HWrgTEG8/VLrCncUyviZabBMjmkMqLQkZAXyRrPOTKiT2kte+1tBSiL8mip2uzCOKq6621t7iQXpG5SfojSJ5GuMKCCMj780iqfNXb0xiyFtZb6Ch+VPDD9fnQI2IdibI7DLPMaHE+hPuLGRpIhcTF/cAYJbhOH8XQp08aj5In3Fo+TR80lsxv5/dknHfkRsj9rDIoA9OtyXnh8C6JMtHwYM6USrEonNQo7nyf515OxC+owqKRIBYEST9lpguse1kpPEeB7Kx0/+MFvBKeXPwmpvsTrlBcKy5tk0TAJHCdce85injIJr9GU0kUWUQlqZerNWm1aCxImMS4m7rnreS1Ul4JfXLWbHFYGMmiET2YTw780bUpuoarZDxwEKqx1AiQLBhklFwB22VMN79n+7ORaLaB9O/r64tlnSwC9ppYPLQ9eGctMZIGYyPD3NhY5Ka6LLmdtslJhDF3Y5mPmYxOW8mRw4jMwEN8H/QELHlk2F01x9GkWhkN39i4LrneBwwKPsBGIJmzYk6sXfAZif2nNkEn7PPlh8aZBsxhzwIgW4HZApW+jBsWWs+YvPjA1D4od5XIse3Lmclj8cvFml0/A/4FiuNqLYKXYDmEaUvw2l08G1pTBhtOLs567NLUo7GiXKuUuXPs8v0pe/XX4ateTLScSlOKL0B9qe8NdiGXEeJLe3ZEfsOcF6+7mZMrSZVvjOl0iEDIDSbraSs4FH3ORkjO+C6J6iG2yLtPpHoi1SMj1YvXS33XQPpHQnrDLJdS619TxwOG+JELUYhwCmDYNgmmZCl9j12g4MMgD/TwEQdJFF7atlK93FapnLATLD7CSVQSCA8Bq9bYDoEZuBNNGJ6o9UStR0Stl2m9OiNkw3J2FdG+1N/Q62QT8m0wih47r4EEwctQlcIJX5vQ1ylaei0OdRsdsnsEP5HqiVSPi1SHXdp2oFBgCUGNjxx5E49UZ+dk/lqnNjSLHVZ5vV1ccctxjhPWhYNvbGyudFP7MnZs8LjifrzZQ4LbZ2+7uim0prFNTZqWcfe2xjpssLA9OM9v42AU6zemmMYu4m8h9G5tW3JIb+nX9h57h4icE52toXVsPPomytGfO/dR8GTZgnggf7RUpT4Gl5FZSdejPcAFZq5iPkZ0bAqikgmHvMCo4+xgcnRggjy6StnnDmZrZJxTCyYU3EjZ2lKeGHtfHR8cYKO7P4ZoUewrxUu+jxyTf+vlpsyIVk4gc0g/DRPfttX1yy29IzJdLmn3bpmnmqSwoMs6/leRMg+VYWgCaJpYfSBb2BXZwq6aDB4ACGnkSFy3yVEHF6nguZXUz9NilEoOqwvGTySE3Gi29kF1jcAQuqiMpJ80uG77/rmxMZsX1EynQm+HPG/799yKG00Z2dgAmE3hAzqdKxSv8INMtOojtgLKBSbRcqOuJtHcnpJXIpDkvnumNJTw4oc/qBSOKeF3JRt2ARFiX2J0I8go6kch9AnQcH+8nvcBC0ITTZ/NBjgIfbZqPqfHwIklGOEy7EYGtM4AVSEdXRcLumz15tllQ47nbNcS0ulmmx/0lDwisuMOl+B+RyUL6WKM3Q0hF/bGUafRqv1rkrWUslFrKP3EGSd5Lip/79jbhO4X51fXGDxo6llJm8TjzPI7Koz4Hyw2IUohi4Zns0xxPaoBKLyPMunzH7SAoBw= +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Updates a vector store. + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/upload-a-file-that-can-be-used-across-various-endpoints.api.mdx b/versioned_docs/version-v0.2.23/api/upload-a-file-that-can-be-used-across-various-endpoints.api.mdx new file mode 100644 index 0000000..67e454f --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/upload-a-file-that-can-be-used-across-various-endpoints.api.mdx @@ -0,0 +1,72 @@ +--- +id: upload-a-file-that-can-be-used-across-various-endpoints +title: "Upload a file that can be used across various endpoints." +description: "Upload a file that can be used across various endpoints." +sidebar_label: "Upload a file that can be used across various endpoints." +hide_title: true +hide_table_of_contents: true +api: eJztWF1v2zYU/SuEnhLAX0nXdTWwh7RrsQItEuRjwBAHBS3RFluJ1EjKqWf4v+9cUpJlW8HcfTwEcF5iy+S55L3nnCtyFRlhC62ssNF4FZ2PRvQvETY2snBSq2gcXSh2WQh18eG9zMTl9IuIHTOiwEShnFRz5lLByiLTPBEJm2HQIOpFsVYOvxMcL4pMxpzghl8sYa4iG6ci5/TJLQuBKNoDY2JhdCGMk2FF1ePNOOsMYoYAFj9EFBBfEzHjZbb1oL2JW6wxYDEC6rHHVMYpk5bx7JEvLZv4eZMoWvcimXQF3AekGUwmlIaZFKYGjbliU4EczYQRKkZSpPJJurj6wIRKCi2VsxRounRhm1UsPBdzYTqDWfkntjDzQBS4R6gBAEixEdyJ5DN3B8LdKfmNOZkL63hesBOAWYGcJvaUzbTBVoRqYrFHblkVgqKJb4VE/f+vaBU8RaLviufisILQyHaOCKEoTaFtJ4BQZR6N7yNurcTCqCioCXdxGj10wdMOFZG8wtwOhVg8SSSN59lVi8UznlnRg9L+KLGthCI2bAfVahpsFXErx600bPaDFTrpMtrQrjz3chMGMBpRqwAFhWSk2rCzNcgSVQfYEm3qhy5ToHTQhlBQTw6pFjyTCUMtc56hpLlI/jsXQHFceYhQfr29vWJhNIt14glQZenv+PPOGCzeD+4xzmyqjWO2zHNulnWhhR/TWAdt2kjwxnMYqg+/+zhrwndcZodGDqMpdKYV9sbSMueqD0okfIq6teZsLccbliL6xgfs8uSyCAQ9ZRfs7vrjtmWVFnRwGpXFbLEQLNeGaE/19JVjfKpL54PbQsSwvZjpOC6NN7qddX2PIKoK19VqktdiuU/TE8mruxibGZ3XXjtg17qcp9kS5cky/WjZ9fu37NVPo1eDYGKwosCMml+g+ibeG56w60DxzXoOZH6lnPPX3cqJMwlVsBTTqYki5RqT1bIGJlQQYS4XsEWe6xJjKLfSO8BRVEdRPSdRnb/exLsF0z8R0ytl2ba0ftclNLGgF4BYCGq0FNSgJ7JM5tIN2FUmOBbkwB4+5xBJhh9N3alePtWprDALFB/pJCnhrSwhYpUKPRbKwDdRpeEoraO0npG0Xrb71QdiNlbObgLbN/EreV3sU75ORjJglyWYIHjuu9JUoPEo7c81SS2v5ojVcUJ8EvgoqqOonpeoRl3RDpCQVwlRjc8t7cYfpPYPknf+pgQV9kddl3K3lXgeG20tI05p8Le5LhhMVHPlEC5biJ1lltBMznIIUxbceBbmmxdU6dLxRPX9vDEjhPYh8AQSD5B0uDylqk83dzkDmlgdN8Pcpw7A27c/NK05vc5gSmNWEy6sDi/MWF1F5ildIfnhgVykovpQPWDvmly3Ae/Bc0jzgf3MJq1T8yTq7YyrLhlo4Kqyi/UABumfImlIETY8Fe6RDO/Fj6MRFJyw85evz0f4fHLGkGNDeXkxYglf2lO620KekS7gU50fAoeR7Tc6WRKRWn7XlGVIG+8n3PHvsTx/sdAh5yBFPEHy4Ez/6ppj/xLhqoLaZe5v/pxTl74qIpWr4+agdX9BhfgcCuZv9JS4nGE5u+tc95onqsyyaP2wB1JVsxOlbgYdMOttl6nuB4tml50rfSr2w3obz5lS4AHokGp8jYDp68hdim/DxdkQBVVc0qeZd4ReFF5HrV9+achXUueK8XCIF+N+CgCR9LOM57yPcsVfB7HOsQ+aB6eRbukn/lJ34/sH+o382Re/KudHms9uaD67qVzaCwzxKXgo6eKs4x6NmprdsnfebjhtZG9f4OxCAmSi2M4fOqjwJ9fGxry8gCINWrMJjp2QA1D9zIzHwgZQbsREketq49vRPviUDC5DgzJ8Xq/qE9p9Zj3/qBDonZSTcIH4j513N0UthR/d/Pm7eeUWTnxzwyLDoZLY42W5qmR8H3QShEz6OauuZH17J8HSmNVqirPpncnWa3qMkpll6A/+BXFKsrwnU0vxNgenIhF/FRgSvQ106t/SSmg4EorHXc2D7C1MuohjUbjW8L13a7KFxpeuLm9uyfarHpXTe/HYNxKP29t8pEWu138Buq4efQ== +sidebar_class_name: "post api-method" +info_path: docs/api/llama-stack-specification +custom_edit_url: null +--- + +import MethodEndpoint from "@theme/ApiExplorer/MethodEndpoint"; +import ParamsDetails from "@theme/ParamsDetails"; +import RequestSchema from "@theme/RequestSchema"; +import StatusCodes from "@theme/StatusCodes"; +import OperationTabs from "@theme/OperationTabs"; +import TabItem from "@theme/TabItem"; +import Heading from "@theme/Heading"; + + + + + + + + + + +Upload a file that can be used across various endpoints. +The file upload should be a multipart form request with: +- file: The File object (not file name) to be uploaded. +- purpose: The intended purpose of the uploaded file. +- expires_after: Optional form values describing expiration for the file. Expected expires_after[anchor] = "created_at", expires_after[seconds] = \{integer\}. Seconds must be between 3600 and 2592000 (1 hour to 30 days). + + + + + + + + + + + + + + + diff --git a/versioned_docs/version-v0.2.23/api/vector-d-bs.tag.mdx b/versioned_docs/version-v0.2.23/api/vector-d-bs.tag.mdx new file mode 100644 index 0000000..82edbd3 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/vector-d-bs.tag.mdx @@ -0,0 +1,19 @@ +--- +id: vector-d-bs +title: "VectorDBs" +description: "VectorDBs" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/api/vector-io.tag.mdx b/versioned_docs/version-v0.2.23/api/vector-io.tag.mdx new file mode 100644 index 0000000..b2b5479 --- /dev/null +++ b/versioned_docs/version-v0.2.23/api/vector-io.tag.mdx @@ -0,0 +1,19 @@ +--- +id: vector-io +title: "VectorIO" +description: "VectorIO" +custom_edit_url: null +--- + + + + + + + +```mdx-code-block +import DocCardList from '@theme/DocCardList'; +import {useCurrentSidebarCategory} from '@docusaurus/theme-common'; + + +``` diff --git a/versioned_docs/version-v0.2.23/building_applications/agent.mdx b/versioned_docs/version-v0.2.23/building_applications/agent.mdx new file mode 100644 index 0000000..33e98ea --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/agent.mdx @@ -0,0 +1,112 @@ +--- +title: Agents +description: Build powerful AI applications with the Llama Stack agent framework +sidebar_label: Agents +sidebar_position: 3 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Agents + +An Agent in Llama Stack is a powerful abstraction that allows you to build complex AI applications. + +The Llama Stack agent framework is built on a modular architecture that allows for flexible and powerful AI applications. This document explains the key components and how they work together. + +## Core Concepts + +### 1. Agent Configuration + +Agents are configured using the `AgentConfig` class, which includes: + +- **Model**: The underlying LLM to power the agent +- **Instructions**: System prompt that defines the agent's behavior +- **Tools**: Capabilities the agent can use to interact with external systems +- **Safety Shields**: Guardrails to ensure responsible AI behavior + +```python +from llama_stack_client import Agent + +# Create the agent +agent = Agent( + llama_stack_client, + model="meta-llama/Llama-3-70b-chat", + instructions="You are a helpful assistant that can use tools to answer questions.", + tools=["builtin::code_interpreter", "builtin::rag/knowledge_search"], +) +``` + +### 2. Sessions + +Agents maintain state through sessions, which represent a conversation thread: + +```python +# Create a session +session_id = agent.create_session(session_name="My conversation") +``` + +### 3. Turns + +Each interaction with an agent is called a "turn" and consists of: + +- **Input Messages**: What the user sends to the agent +- **Steps**: The agent's internal processing (inference, tool execution, etc.) +- **Output Message**: The agent's response + + + + +```python +from llama_stack_client import AgentEventLogger + +# Create a turn with streaming response +turn_response = agent.create_turn( + session_id=session_id, + messages=[{"role": "user", "content": "Tell me about Llama models"}], +) +for log in AgentEventLogger().log(turn_response): + log.print() +``` + + + + +```python +from rich.pretty import pprint + +# Non-streaming API +response = agent.create_turn( + session_id=session_id, + messages=[{"role": "user", "content": "Tell me about Llama models"}], + stream=False, +) +print("Inputs:") +pprint(response.input_messages) +print("Output:") +pprint(response.output_message.content) +print("Steps:") +pprint(response.steps) +``` + + + + +### 4. Steps + +Each turn consists of multiple steps that represent the agent's thought process: + +- **Inference Steps**: The agent generating text responses +- **Tool Execution Steps**: The agent using tools to gather information +- **Shield Call Steps**: Safety checks being performed + +## Agent Execution Loop + +Refer to the [Agent Execution Loop](./agent_execution_loop) for more details on what happens within an agent turn. + +## Related Resources + +- **[Agent Execution Loop](./agent_execution_loop)** - Understanding the internal processing flow +- **[RAG (Retrieval Augmented Generation)](./rag)** - Building knowledge-enhanced agents +- **[Tools Integration](./tools)** - Extending agent capabilities with external tools +- **[Safety Guardrails](./safety)** - Implementing responsible AI practices diff --git a/versioned_docs/version-v0.2.23/building_applications/agent_execution_loop.mdx b/versioned_docs/version-v0.2.23/building_applications/agent_execution_loop.mdx new file mode 100644 index 0000000..458e997 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/agent_execution_loop.mdx @@ -0,0 +1,185 @@ +--- +title: Agent Execution Loop +description: Understanding the internal processing flow of Llama Stack agents +sidebar_label: Agent Execution Loop +sidebar_position: 4 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Agent Execution Loop + +Agents are the heart of Llama Stack applications. They combine inference, memory, safety, and tool usage into coherent workflows. At its core, an agent follows a sophisticated execution loop that enables multi-step reasoning, tool usage, and safety checks. + +## Steps in the Agent Workflow + +Each agent turn follows these key steps: + +1. **Initial Safety Check**: The user's input is first screened through configured safety shields + +2. **Context Retrieval**: + - If RAG is enabled, the agent can choose to query relevant documents from memory banks. You can use the `instructions` field to steer the agent. + - For new documents, they are first inserted into the memory bank. + - Retrieved context is provided to the LLM as a tool response in the message history. + +3. **Inference Loop**: The agent enters its main execution loop: + - The LLM receives a user prompt (with previous tool outputs) + - The LLM generates a response, potentially with [tool calls](./tools) + - If tool calls are present: + - Tool inputs are safety-checked + - Tools are executed (e.g., web search, code execution) + - Tool responses are fed back to the LLM for synthesis + - The loop continues until: + - The LLM provides a final response without tool calls + - Maximum iterations are reached + - Token limit is exceeded + +4. **Final Safety Check**: The agent's final response is screened through safety shields + +## Execution Flow Diagram + +```mermaid +sequenceDiagram + participant U as User + participant E as Executor + participant M as Memory Bank + participant L as LLM + participant T as Tools + participant S as Safety Shield + + Note over U,S: Agent Turn Start + U->>S: 1. Submit Prompt + activate S + S->>E: Input Safety Check + deactivate S + + loop Inference Loop + E->>L: 2.1 Augment with Context + L-->>E: 2.2 Response (with/without tool calls) + + alt Has Tool Calls + E->>S: Check Tool Input + S->>T: 3.1 Execute Tool + T-->>E: 3.2 Tool Response + E->>L: 4.1 Tool Response + L-->>E: 4.2 Synthesized Response + end + + opt Stop Conditions + Note over E: Break if: + Note over E: - No tool calls + Note over E: - Max iterations reached + Note over E: - Token limit exceeded + end + end + + E->>S: Output Safety Check + S->>U: 5. Final Response +``` + +Each step in this process can be monitored and controlled through configurations. + +## Agent Execution Example + +Here's an example that demonstrates monitoring the agent's execution: + + + + +```python +from llama_stack_client import LlamaStackClient, Agent, AgentEventLogger + +# Replace host and port +client = LlamaStackClient(base_url=f"http://{HOST}:{PORT}") + +agent = Agent( + client, + # Check with `llama-stack-client models list` + model="Llama3.2-3B-Instruct", + instructions="You are a helpful assistant", + # Enable both RAG and tool usage + tools=[ + { + "name": "builtin::rag/knowledge_search", + "args": {"vector_db_ids": ["my_docs"]}, + }, + "builtin::code_interpreter", + ], + # Configure safety (optional) + input_shields=["llama_guard"], + output_shields=["llama_guard"], + # Control the inference loop + max_infer_iters=5, + sampling_params={ + "strategy": {"type": "top_p", "temperature": 0.7, "top_p": 0.95}, + "max_tokens": 2048, + }, +) +session_id = agent.create_session("monitored_session") + +# Stream the agent's execution steps +response = agent.create_turn( + messages=[{"role": "user", "content": "Analyze this code and run it"}], + documents=[ + { + "content": "https://raw.githubusercontent.com/example/code.py", + "mime_type": "text/plain", + } + ], + session_id=session_id, +) + +# Monitor each step of execution +for log in AgentEventLogger().log(response): + log.print() +``` + + + + +```python +from rich.pretty import pprint + +# Using non-streaming API, the response contains input, steps, and output. +response = agent.create_turn( + messages=[{"role": "user", "content": "Analyze this code and run it"}], + documents=[ + { + "content": "https://raw.githubusercontent.com/example/code.py", + "mime_type": "text/plain", + } + ], + session_id=session_id, + stream=False, +) + +pprint(f"Input: {response.input_messages}") +pprint(f"Output: {response.output_message.content}") +pprint(f"Steps: {response.steps}") +``` + + + + +## Key Configuration Options + +### Loop Control +- **max_infer_iters**: Maximum number of inference iterations (default: 5) +- **max_tokens**: Token limit for responses +- **temperature**: Controls response randomness + +### Safety Configuration +- **input_shields**: Safety checks for user input +- **output_shields**: Safety checks for agent responses + +### Tool Integration +- **tools**: List of available tools for the agent +- **tool_choice**: Control over when tools are used + +## Related Resources + +- **[Agents](./agent)** - Understanding agent fundamentals +- **[Tools Integration](./tools)** - Adding capabilities to agents +- **[Safety Guardrails](./safety)** - Implementing safety measures +- **[RAG (Retrieval Augmented Generation)](./rag)** - Building knowledge-enhanced workflows diff --git a/versioned_docs/version-v0.2.23/building_applications/evals.mdx b/versioned_docs/version-v0.2.23/building_applications/evals.mdx new file mode 100644 index 0000000..d2eb0bd --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/evals.mdx @@ -0,0 +1,256 @@ +--- +title: Evaluations +description: Evaluate LLM applications with Llama Stack's comprehensive evaluation framework +sidebar_label: Evaluations +sidebar_position: 7 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +This guide walks you through the process of evaluating an LLM application built using Llama Stack. For detailed API reference, check out the [Evaluation Reference](../references/evals_reference/) guide that covers the complete set of APIs and developer experience flow. + +:::tip[Interactive Examples] +Check out our [Colab notebook](https://colab.research.google.com/drive/10CHyykee9j2OigaIcRv47BKG9mrNm0tJ?usp=sharing) for working examples with evaluations, or try the [Getting Started notebook](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb). +::: + +## Application Evaluation Example + +[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb) + +Llama Stack offers a library of scoring functions and the `/scoring` API, allowing you to run evaluations on your pre-annotated AI application datasets. + +In this example, we will show you how to: +1. **Build an Agent** with Llama Stack +2. **Query the agent's sessions, turns, and steps** to analyze execution +3. **Evaluate the results** using scoring functions + +## Step-by-Step Evaluation Process + +### 1. Building a Search Agent + +First, let's create an agent that can search the web to answer questions: + +```python +from llama_stack_client import LlamaStackClient, Agent, AgentEventLogger + +client = LlamaStackClient(base_url=f"http://{HOST}:{PORT}") + +agent = Agent( + client, + model="meta-llama/Llama-3.3-70B-Instruct", + instructions="You are a helpful assistant. Use search tool to answer the questions.", + tools=["builtin::websearch"], +) + +# Test prompts for evaluation +user_prompts = [ + "Which teams played in the NBA Western Conference Finals of 2024. Search the web for the answer.", + "In which episode and season of South Park does Bill Cosby (BSM-471) first appear? Give me the number and title. Search the web for the answer.", + "What is the British-American kickboxer Andrew Tate's kickboxing name? Search the web for the answer.", +] + +session_id = agent.create_session("test-session") + +# Execute all prompts in the session +for prompt in user_prompts: + response = agent.create_turn( + messages=[ + { + "role": "user", + "content": prompt, + } + ], + session_id=session_id, + ) + + for log in AgentEventLogger().log(response): + log.print() +``` + +### 2. Query Agent Execution Steps + +Now, let's analyze the agent's execution steps to understand its performance: + + + + +```python +from rich.pretty import pprint + +# Query the agent's session to get detailed execution data +session_response = client.agents.session.retrieve( + session_id=session_id, + agent_id=agent.agent_id, +) + +pprint(session_response) +``` + + + + +```python +# Sanity check: Verify that all user prompts are followed by tool calls +num_tool_call = 0 +for turn in session_response.turns: + for step in turn.steps: + if ( + step.step_type == "tool_execution" + and step.tool_calls[0].tool_name == "brave_search" + ): + num_tool_call += 1 + +print( + f"{num_tool_call}/{len(session_response.turns)} user prompts are followed by a tool call to `brave_search`" +) +``` + + + + +### 3. Evaluate Agent Responses + +Now we'll evaluate the agent's responses using Llama Stack's scoring API: + + + + +```python +# Process agent execution history into evaluation rows +eval_rows = [] + +# Define expected answers for our test prompts +expected_answers = [ + "Dallas Mavericks and the Minnesota Timberwolves", + "Season 4, Episode 12", + "King Cobra", +] + +# Create evaluation dataset from agent responses +for i, turn in enumerate(session_response.turns): + eval_rows.append( + { + "input_query": turn.input_messages[0].content, + "generated_answer": turn.output_message.content, + "expected_answer": expected_answers[i], + } + ) + +pprint(eval_rows) +``` + + + + +```python +# Configure scoring parameters +scoring_params = { + "basic::subset_of": None, # Check if generated answer contains expected answer +} + +# Run evaluation using Llama Stack's scoring API +scoring_response = client.scoring.score( + input_rows=eval_rows, + scoring_functions=scoring_params +) + +pprint(scoring_response) + +# Analyze results +for i, result in enumerate(scoring_response.results): + print(f"Query {i+1}: {result.score}") + print(f" Generated: {eval_rows[i]['generated_answer'][:100]}...") + print(f" Expected: {expected_answers[i]}") + print(f" Score: {result.score}") + print() +``` + + + + +## Available Scoring Functions + +Llama Stack provides several built-in scoring functions: + +### Basic Scoring Functions +- **`basic::subset_of`**: Checks if the expected answer is contained in the generated response +- **`basic::exact_match`**: Performs exact string matching between expected and generated answers +- **`basic::regex_match`**: Uses regular expressions to match patterns in responses + +### Advanced Scoring Functions +- **`llm_as_judge::accuracy`**: Uses an LLM to judge response accuracy +- **`llm_as_judge::helpfulness`**: Evaluates how helpful the response is +- **`llm_as_judge::safety`**: Assesses response safety and appropriateness + +### Custom Scoring Functions +You can also create custom scoring functions for domain-specific evaluation needs. + +## Evaluation Workflow Best Practices + +### ๐ŸŽฏ **Dataset Preparation** +- Use diverse test cases that cover edge cases and common scenarios +- Include clear expected answers or success criteria +- Balance your dataset across different difficulty levels + +### ๐Ÿ“Š **Metrics Selection** +- Choose appropriate scoring functions for your use case +- Combine multiple metrics for comprehensive evaluation +- Consider both automated and human evaluation metrics + +### ๐Ÿ”„ **Iterative Improvement** +- Run evaluations regularly during development +- Use evaluation results to identify areas for improvement +- Track performance changes over time + +### ๐Ÿ“ˆ **Analysis & Reporting** +- Analyze failures to understand model limitations +- Generate comprehensive evaluation reports +- Share results with stakeholders for informed decision-making + +## Advanced Evaluation Scenarios + +### Batch Evaluation +For evaluating large datasets efficiently: + +```python +# Prepare large evaluation dataset +large_eval_dataset = [ + {"input_query": query, "expected_answer": answer} + for query, answer in zip(queries, expected_answers) +] + +# Run batch evaluation +batch_results = client.scoring.score( + input_rows=large_eval_dataset, + scoring_functions={ + "basic::subset_of": None, + "llm_as_judge::accuracy": {"judge_model": "meta-llama/Llama-3.3-70B-Instruct"}, + } +) +``` + +### Multi-Metric Evaluation +Combining different scoring approaches: + +```python +comprehensive_scoring = { + "exact_match": "basic::exact_match", + "subset_match": "basic::subset_of", + "llm_judge": "llm_as_judge::accuracy", + "safety_check": "llm_as_judge::safety", +} + +results = client.scoring.score( + input_rows=eval_rows, + scoring_functions=comprehensive_scoring +) +``` + +## Related Resources + +- **[Agents](./agent)** - Building agents for evaluation +- **[Tools Integration](./tools)** - Using tools in evaluated agents +- **[Evaluation Reference](../references/evals_reference/)** - Complete API reference for evaluations +- **[Getting Started Notebook](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb)** - Interactive examples +- **[Evaluation Examples](https://colab.research.google.com/drive/10CHyykee9j2OigaIcRv47BKG9mrNm0tJ?usp=sharing)** - Additional evaluation scenarios diff --git a/versioned_docs/version-v0.2.23/building_applications/index.mdx b/versioned_docs/version-v0.2.23/building_applications/index.mdx new file mode 100644 index 0000000..a4b71ef --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/index.mdx @@ -0,0 +1,83 @@ +--- +title: Building Applications +description: Comprehensive guides for building AI applications with Llama Stack +sidebar_label: Overview +sidebar_position: 5 +--- + +# AI Application Examples + +Llama Stack provides all the building blocks needed to create sophisticated AI applications. + +## Getting Started + +The best way to get started is to look at this comprehensive notebook which walks through the various APIs (from basic inference, to RAG agents) and how to use them. + +**๐Ÿ““ [Building AI Applications Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb)** + +## Core Topics + +Here are the key topics that will help you build effective AI applications: + +### ๐Ÿค– **Agent Development** +- **[Agent Framework](./agent.mdx)** - Understand the components and design patterns of the Llama Stack agent framework +- **[Agent Execution Loop](./agent_execution_loop.mdx)** - How agents process information, make decisions, and execute actions +- **[Agents vs Responses API](./responses_vs_agents.mdx)** - Learn when to use each API for different use cases + +### ๐Ÿ“š **Knowledge Integration** +- **[RAG (Retrieval-Augmented Generation)](./rag.mdx)** - Enhance your agents with external knowledge through retrieval mechanisms + +### ๐Ÿ› ๏ธ **Capabilities & Extensions** +- **[Tools](./tools.mdx)** - Extend your agents' capabilities by integrating with external tools and APIs + +### ๐Ÿ“Š **Quality & Monitoring** +- **[Evaluations](./evals.mdx)** - Evaluate your agents' effectiveness and identify areas for improvement +- **[Telemetry](./telemetry.mdx)** - Monitor and analyze your agents' performance and behavior +- **[Safety](./safety.mdx)** - Implement guardrails and safety measures to ensure responsible AI behavior + +### ๐ŸŽฎ **Interactive Development** +- **[Playground](./playground.mdx)** - Interactive environment for testing and developing applications + +## Application Patterns + +### ๐Ÿค– **Conversational Agents** +Build intelligent chatbots and assistants that can: +- Maintain context across conversations +- Access external knowledge bases +- Execute actions through tool integrations +- Apply safety filters and guardrails + +### ๐Ÿ“– **RAG Applications** +Create knowledge-augmented applications that: +- Retrieve relevant information from documents +- Generate contextually accurate responses +- Handle large knowledge bases efficiently +- Provide source attribution + +### ๐Ÿ”ง **Tool-Enhanced Systems** +Develop applications that can: +- Search the web for real-time information +- Interact with databases and APIs +- Perform calculations and analysis +- Execute complex multi-step workflows + +### ๐Ÿ›ก๏ธ **Enterprise Applications** +Build production-ready systems with: +- Comprehensive safety measures +- Performance monitoring and analytics +- Scalable deployment configurations +- Evaluation and quality assurance + +## Next Steps + +1. **๐Ÿ“– Start with the Notebook** - Work through the complete tutorial +2. **๐ŸŽฏ Choose Your Pattern** - Pick the application type that matches your needs +3. **๐Ÿ—๏ธ Build Your Foundation** - Set up your [providers](/docs/providers/) and [distributions](/docs/distributions/) +4. **๐Ÿš€ Deploy & Monitor** - Use our [deployment guides](/docs/deploying/) for production + +## Related Resources + +- **[Getting Started](/docs/getting_started/quickstart)** - Basic setup and concepts +- **[Providers](/docs/providers/)** - Available AI service providers +- **[Distributions](/docs/distributions/)** - Pre-configured deployment packages +- **[API Reference](/docs/api/llama-stack-specification)** - Complete API documentation diff --git a/versioned_docs/version-v0.2.23/building_applications/playground.mdx b/versioned_docs/version-v0.2.23/building_applications/playground.mdx new file mode 100644 index 0000000..b2aa1b4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/playground.mdx @@ -0,0 +1,299 @@ +--- +title: Llama Stack Playground +description: Interactive interface to explore and experiment with Llama Stack capabilities +sidebar_label: Playground +sidebar_position: 10 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Llama Stack Playground + +:::note[Experimental Feature] +The Llama Stack Playground is currently experimental and subject to change. We welcome feedback and contributions to help improve it. +::: + +The Llama Stack Playground is a simple interface that aims to: +- **Showcase capabilities and concepts** of Llama Stack in an interactive environment +- **Demo end-to-end application code** to help users get started building their own applications +- **Provide a UI** to help users inspect and understand Llama Stack API providers and resources + +## Key Features + +### Interactive Playground Pages + +The playground provides interactive pages for users to explore Llama Stack API capabilities: + +#### Chatbot Interface + + + + + + +**Simple Chat Interface** +- Chat directly with Llama models through an intuitive interface +- Uses the `/inference/chat-completion` streaming API under the hood +- Real-time message streaming for responsive interactions +- Perfect for testing model capabilities and prompt engineering + + + + +**Document-Aware Conversations** +- Upload documents to create memory banks +- Chat with a RAG-enabled agent that can query your documents +- Uses Llama Stack's `/agents` API to create and manage RAG sessions +- Ideal for exploring knowledge-enhanced AI applications + + + + +#### Evaluation Interface + + + + + + +**Custom Dataset Evaluation** +- Upload your own evaluation datasets +- Run evaluations using available scoring functions +- Uses Llama Stack's `/scoring` API for flexible evaluation workflows +- Great for testing application performance on custom metrics + + + + + + +**Pre-registered Evaluation Tasks** +- Evaluate models or agents on pre-defined tasks +- Uses Llama Stack's `/eval` API for comprehensive evaluation +- Combines datasets and scoring functions for standardized testing + +**Setup Requirements:** +Register evaluation datasets and benchmarks first: + +```bash +# Register evaluation dataset +llama-stack-client datasets register \ + --dataset-id "mmlu" \ + --provider-id "huggingface" \ + --url "https://huggingface.co/datasets/llamastack/evals" \ + --metadata '{"path": "llamastack/evals", "name": "evals__mmlu__details", "split": "train"}' \ + --schema '{"input_query": {"type": "string"}, "expected_answer": {"type": "string"}, "chat_completion_input": {"type": "string"}}' + +# Register benchmark task +llama-stack-client benchmarks register \ + --eval-task-id meta-reference-mmlu \ + --provider-id meta-reference \ + --dataset-id mmlu \ + --scoring-functions basic::regex_parser_multiple_choice_answer +``` + + + + +#### Inspection Interface + + + + + + +**Provider Management** +- Inspect available Llama Stack API providers +- View provider configurations and capabilities +- Uses the `/providers` API for real-time provider information +- Essential for understanding your deployment's capabilities + + + + +**Resource Exploration** +- Inspect Llama Stack API resources including: + - **Models**: Available language models + - **Datasets**: Registered evaluation datasets + - **Memory Banks**: Vector databases and knowledge stores + - **Benchmarks**: Evaluation tasks and scoring functions + - **Shields**: Safety and content moderation tools +- Uses `//list` APIs for comprehensive resource visibility +- For detailed information about resources, see [Core Concepts](/docs/concepts) + + + + +## Getting Started + +### Quick Start Guide + + + + +**1. Start the Llama Stack API Server** + +```bash +# Build and run a distribution (example: together) +llama stack build --distro together --image-type venv +llama stack run together +``` + +**2. Start the Streamlit UI** + +```bash +# Launch the playground interface +uv run --with ".[ui]" streamlit run llama_stack.core/ui/app.py +``` + + + + +**Making the Most of the Playground:** + +- **Start with Chat**: Test basic model interactions and prompt engineering +- **Explore RAG**: Upload sample documents to see knowledge-enhanced responses +- **Try Evaluations**: Use the scoring interface to understand evaluation metrics +- **Inspect Resources**: Check what providers and resources are available +- **Experiment with Settings**: Adjust parameters to see how they affect results + + + + +### Available Distributions + +The playground works with any Llama Stack distribution. Popular options include: + + + + +```bash +llama stack build --distro together --image-type venv +llama stack run together +``` + +**Features:** +- Cloud-hosted models +- Fast inference +- Multiple model options + + + + +```bash +llama stack build --distro ollama --image-type venv +llama stack run ollama +``` + +**Features:** +- Local model execution +- Privacy-focused +- No internet required + + + + +```bash +llama stack build --distro meta-reference --image-type venv +llama stack run meta-reference +``` + +**Features:** +- Reference implementation +- All API features available +- Best for development + + + + +## Use Cases & Examples + +### Educational Use Cases +- **Learning Llama Stack**: Hands-on exploration of API capabilities +- **Prompt Engineering**: Interactive testing of different prompting strategies +- **RAG Experimentation**: Understanding how document retrieval affects responses +- **Evaluation Understanding**: See how different metrics evaluate model performance + +### Development Use Cases +- **Prototype Testing**: Quick validation of application concepts +- **API Exploration**: Understanding available endpoints and parameters +- **Integration Planning**: Seeing how different components work together +- **Demo Creation**: Showcasing Llama Stack capabilities to stakeholders + +### Research Use Cases +- **Model Comparison**: Side-by-side testing of different models +- **Evaluation Design**: Understanding how scoring functions work +- **Safety Testing**: Exploring shield effectiveness with different inputs +- **Performance Analysis**: Measuring model behavior across different scenarios + +## Best Practices + +### ๐Ÿš€ **Getting Started** +- Begin with simple chat interactions to understand basic functionality +- Gradually explore more advanced features like RAG and evaluations +- Use the inspection tools to understand your deployment's capabilities + +### ๐Ÿ”ง **Development Workflow** +- Use the playground to prototype before writing application code +- Test different parameter settings interactively +- Validate evaluation approaches before implementing them programmatically + +### ๐Ÿ“Š **Evaluation & Testing** +- Start with simple scoring functions before trying complex evaluations +- Use the playground to understand evaluation results before automation +- Test safety features with various input types + +### ๐ŸŽฏ **Production Preparation** +- Use playground insights to inform your production API usage +- Test edge cases and error conditions interactively +- Validate resource configurations before deployment + +## Related Resources + +- **[Getting Started Guide](../getting_started/quickstart)** - Complete setup and introduction +- **[Core Concepts](/docs/concepts)** - Understanding Llama Stack fundamentals +- **[Agents](./agent)** - Building intelligent agents +- **[RAG (Retrieval Augmented Generation)](./rag)** - Knowledge-enhanced applications +- **[Evaluations](./evals)** - Comprehensive evaluation framework +- **[API Reference](/docs/api/llama-stack-specification)** - Complete API documentation diff --git a/versioned_docs/version-v0.2.23/building_applications/rag.mdx b/versioned_docs/version-v0.2.23/building_applications/rag.mdx new file mode 100644 index 0000000..5212616 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/rag.mdx @@ -0,0 +1,367 @@ +--- +title: Retrieval Augmented Generation (RAG) +description: Build knowledge-enhanced AI applications with external document retrieval +sidebar_label: RAG (Retrieval Augmented Generation) +sidebar_position: 2 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Retrieval Augmented Generation (RAG) + +RAG enables your applications to reference and recall information from previous interactions or external documents. + +## Architecture Overview + +Llama Stack organizes the APIs that enable RAG into three layers: + +1. **Lower-Level APIs**: Deal with raw storage and retrieval. These include Vector IO, KeyValue IO (coming soon) and Relational IO (also coming soon) +2. **RAG Tool**: A first-class tool as part of the [Tools API](./tools) that allows you to ingest documents (from URLs, files, etc) with various chunking strategies and query them smartly +3. **Agents API**: The top-level [Agents API](./agent) that allows you to create agents that can use the tools to answer questions, perform tasks, and more + +![RAG System Architecture](/img/rag.png) + +The RAG system uses lower-level storage for different types of data: +- **Vector IO**: For semantic search and retrieval +- **Key-Value and Relational IO**: For structured data storage + +:::info[Future Storage Types] +We may add more storage types like Graph IO in the future. +::: + +## Setting up Vector Databases + +For this guide, we will use [Ollama](https://ollama.com/) as the inference provider. Ollama is an LLM runtime that allows you to run Llama models locally. + +Here's how to set up a vector database for RAG: + +```python +# Create HTTP client +import os +from llama_stack_client import LlamaStackClient + +client = LlamaStackClient(base_url=f"http://localhost:{os.environ['LLAMA_STACK_PORT']}") + +# Register a vector database +vector_db_id = "my_documents" +response = client.vector_dbs.register( + vector_db_id=vector_db_id, + embedding_model="all-MiniLM-L6-v2", + embedding_dimension=384, + provider_id="faiss", +) +``` + +## Document Ingestion + +You can ingest documents into the vector database using two methods: directly inserting pre-chunked documents or using the RAG Tool. + +### Direct Document Insertion + + + + +```python +# You can insert a pre-chunked document directly into the vector db +chunks = [ + { + "content": "Your document text here", + "mime_type": "text/plain", + "metadata": { + "document_id": "doc1", + "author": "Jane Doe", + }, + }, +] +client.vector_io.insert(vector_db_id=vector_db_id, chunks=chunks) +``` + + + + +If you decide to precompute embeddings for your documents, you can insert them directly into the vector database by including the embedding vectors in the chunk data. This is useful if you have a separate embedding service or if you want to customize the ingestion process. + +```python +chunks_with_embeddings = [ + { + "content": "First chunk of text", + "mime_type": "text/plain", + "embedding": [0.1, 0.2, 0.3, ...], # Your precomputed embedding vector + "metadata": {"document_id": "doc1", "section": "introduction"}, + }, + { + "content": "Second chunk of text", + "mime_type": "text/plain", + "embedding": [0.2, 0.3, 0.4, ...], # Your precomputed embedding vector + "metadata": {"document_id": "doc1", "section": "methodology"}, + }, +] +client.vector_io.insert(vector_db_id=vector_db_id, chunks=chunks_with_embeddings) +``` + +:::warning[Embedding Dimensions] +When providing precomputed embeddings, ensure the embedding dimension matches the `embedding_dimension` specified when registering the vector database. +::: + + + + +### Document Retrieval + +You can query the vector database to retrieve documents based on their embeddings. + +```python +# You can then query for these chunks +chunks_response = client.vector_io.query( + vector_db_id=vector_db_id, + query="What do you know about..." +) +``` + +## Using the RAG Tool + +:::danger[Deprecation Notice] +The RAG Tool is being deprecated in favor of directly using the OpenAI-compatible Search API. We recommend migrating to the OpenAI APIs for better compatibility and future support. +::: + +A better way to ingest documents is to use the RAG Tool. This tool allows you to ingest documents from URLs, files, etc. and automatically chunks them into smaller pieces. More examples for how to format a RAGDocument can be found in the [appendix](#more-ragdocument-examples). + +### OpenAI API Integration & Migration + +The RAG tool has been updated to use OpenAI-compatible APIs. This provides several benefits: + +- **Files API Integration**: Documents are now uploaded using OpenAI's file upload endpoints +- **Vector Stores API**: Vector storage operations use OpenAI's vector store format with configurable chunking strategies +- **Error Resilience**: When processing multiple documents, individual failures are logged but don't crash the operation. Failed documents are skipped while successful ones continue processing. + +### Migration Path + +We recommend migrating to the OpenAI-compatible Search API for: + +1. **Better OpenAI Ecosystem Integration**: Direct compatibility with OpenAI tools and workflows including the Responses API +2. **Future-Proof**: Continued support and feature development +3. **Full OpenAI Compatibility**: Vector Stores, Files, and Search APIs are fully compatible with OpenAI's Responses API + +The OpenAI APIs are used under the hood, so you can continue to use your existing RAG Tool code with minimal changes. However, we recommend updating your code to use the new OpenAI-compatible APIs for better long-term support. If any documents fail to process, they will be logged in the response but will not cause the entire operation to fail. + +### RAG Tool Example + +```python +from llama_stack_client import RAGDocument + +urls = ["memory_optimizations.rst", "chat.rst", "llama3.rst"] +documents = [ + RAGDocument( + document_id=f"num-{i}", + content=f"https://raw.githubusercontent.com/pytorch/torchtune/main/docs/source/tutorials/{url}", + mime_type="text/plain", + metadata={}, + ) + for i, url in enumerate(urls) +] + +client.tool_runtime.rag_tool.insert( + documents=documents, + vector_db_id=vector_db_id, + chunk_size_in_tokens=512, +) + +# Query documents +results = client.tool_runtime.rag_tool.query( + vector_db_ids=[vector_db_id], + content="What do you know about...", +) +``` + +### Custom Context Configuration + +You can configure how the RAG tool adds metadata to the context if you find it useful for your application: + +```python +# Query documents with custom template +results = client.tool_runtime.rag_tool.query( + vector_db_ids=[vector_db_id], + content="What do you know about...", + query_config={ + "chunk_template": "Result {index}\nContent: {chunk.content}\nMetadata: {metadata}\n", + }, +) +``` + +## Building RAG-Enhanced Agents + +One of the most powerful patterns is combining agents with RAG capabilities. Here's a complete example: + +### Agent with Knowledge Search + +```python +from llama_stack_client import Agent + +# Create agent with memory +agent = Agent( + client, + model="meta-llama/Llama-3.3-70B-Instruct", + instructions="You are a helpful assistant", + tools=[ + { + "name": "builtin::rag/knowledge_search", + "args": { + "vector_db_ids": [vector_db_id], + # Defaults + "query_config": { + "chunk_size_in_tokens": 512, + "chunk_overlap_in_tokens": 0, + "chunk_template": "Result {index}\nContent: {chunk.content}\nMetadata: {metadata}\n", + }, + }, + } + ], +) +session_id = agent.create_session("rag_session") + +# Ask questions about documents in the vector db, and the agent will query the db to answer the question. +response = agent.create_turn( + messages=[{"role": "user", "content": "How to optimize memory in PyTorch?"}], + session_id=session_id, +) +``` + +:::tip[Agent Instructions] +The `instructions` field in the `AgentConfig` can be used to guide the agent's behavior. It is important to experiment with different instructions to see what works best for your use case. +::: + +### Document-Aware Conversations + +You can also pass documents along with the user's message and ask questions about them: + +```python +# Initial document ingestion +response = agent.create_turn( + messages=[ + {"role": "user", "content": "I am providing some documents for reference."} + ], + documents=[ + { + "content": "https://raw.githubusercontent.com/pytorch/torchtune/main/docs/source/tutorials/memory_optimizations.rst", + "mime_type": "text/plain", + } + ], + session_id=session_id, +) + +# Query with RAG +response = agent.create_turn( + messages=[{"role": "user", "content": "What are the key topics in the documents?"}], + session_id=session_id, +) +``` + +### Viewing Agent Responses + +You can print the response with the following: + +```python +from llama_stack_client import AgentEventLogger + +for log in AgentEventLogger().log(response): + log.print() +``` + +## Vector Database Management + +### Unregistering Vector DBs + +If you need to clean up and unregister vector databases, you can do so as follows: + + + + +```python +# Unregister a specified vector database +vector_db_id = "my_vector_db_id" +print(f"Unregistering vector database: {vector_db_id}") +client.vector_dbs.unregister(vector_db_id=vector_db_id) +``` + + + + +```python +# Unregister all vector databases +for vector_db_id in client.vector_dbs.list(): + print(f"Unregistering vector database: {vector_db_id.identifier}") + client.vector_dbs.unregister(vector_db_id=vector_db_id.identifier) +``` + + + + +## Best Practices + +### ๐ŸŽฏ **Document Chunking** +- Use appropriate chunk sizes (512 tokens is often a good starting point) +- Consider overlap between chunks for better context preservation +- Experiment with different chunking strategies for your content type + +### ๐Ÿ” **Embedding Strategy** +- Choose embedding models that match your domain +- Consider the trade-off between embedding dimension and performance +- Test different embedding models for your specific use case + +### ๐Ÿ“Š **Query Optimization** +- Use specific, well-formed queries for better retrieval +- Experiment with different search strategies +- Consider hybrid approaches (keyword + semantic search) + +### ๐Ÿ›ก๏ธ **Error Handling** +- Implement proper error handling for failed document processing +- Monitor ingestion success rates +- Have fallback strategies for retrieval failures + +## Appendix + +### More RAGDocument Examples + +Here are various ways to create RAGDocument objects for different content types: + +```python +from llama_stack_client import RAGDocument +import base64 + +# File URI +RAGDocument(document_id="num-0", content={"uri": "file://path/to/file"}) + +# Plain text +RAGDocument(document_id="num-1", content="plain text") + +# Explicit text input +RAGDocument( + document_id="num-2", + content={ + "type": "text", + "text": "plain text input", + }, # for inputs that should be treated as text explicitly +) + +# Image from URL +RAGDocument( + document_id="num-3", + content={ + "type": "image", + "image": {"url": {"uri": "https://mywebsite.com/image.jpg"}}, + }, +) + +# Base64 encoded image +B64_ENCODED_IMAGE = base64.b64encode( + requests.get( + "https://raw.githubusercontent.com/meta-llama/llama-stack/refs/heads/main/docs/_static/llama-stack.png" + ).content +) +RAGDocument( + document_id="num-4", + content={"type": "image", "image": {"data": B64_ENCODED_IMAGE}}, +) +``` +For more strongly typed interaction use the typed dicts found [here](https://github.com/meta-llama/llama-stack-client-python/blob/38cd91c9e396f2be0bec1ee96a19771582ba6f17/src/llama_stack_client/types/shared_params/document.py). diff --git a/versioned_docs/version-v0.2.23/building_applications/responses_vs_agents.mdx b/versioned_docs/version-v0.2.23/building_applications/responses_vs_agents.mdx new file mode 100644 index 0000000..4cf7620 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/responses_vs_agents.mdx @@ -0,0 +1,221 @@ +--- +title: Agents vs OpenAI Responses API +description: Compare the Agents API and OpenAI Responses API for building AI applications with tool calling capabilities +sidebar_label: Agents vs Responses API +sidebar_position: 5 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Agents vs OpenAI Responses API + +Llama Stack (LLS) provides two different APIs for building AI applications with tool calling capabilities: the **Agents API** and the **OpenAI Responses API**. While both enable AI systems to use tools, and maintain full conversation history, they serve different use cases and have distinct characteristics. + +:::note +**Note:** For simple and basic inferencing, you may want to use the [Chat Completions API](../providers/openai#chat-completions) directly, before progressing to Agents or Responses API. +::: + +## Overview + +### LLS Agents API +The Agents API is a full-featured, stateful system designed for complex, multi-turn conversations. It maintains conversation state through persistent sessions identified by a unique session ID. The API supports comprehensive agent lifecycle management, detailed execution tracking, and rich metadata about each interaction through a structured session/turn/step hierarchy. The API can orchestrate multiple tool calls within a single turn. + +### OpenAI Responses API +The OpenAI Responses API is a full-featured, stateful system designed for complex, multi-turn conversations, with direct compatibility with OpenAI's conversational patterns enhanced by LLama Stack's tool calling capabilities. It maintains conversation state by chaining responses through a `previous_response_id`, allowing interactions to branch or continue from any prior point. Each response can perform multiple tool calls within a single turn. + +### Key Differences +The LLS Agents API uses the Chat Completions API on the backend for inference as it's the industry standard for building AI applications and most LLM providers are compatible with this API. For a detailed comparison between Responses and Chat Completions, see [OpenAI's documentation](https://platform.openai.com/docs/guides/responses-vs-chat-completions). + +Additionally, Agents let you specify input/output shields whereas Responses do not (though support is planned). Agents use a linear conversation model referenced by a single session ID. Responses, on the other hand, support branching, where each response can serve as a fork point, and conversations are tracked by the latest response ID. Responses also lets you dynamically choose the model, vector store, files, MCP servers, and more on each inference call, enabling more complex workflows. Agents require a static configuration for these components at the start of the session. + +Today the Agents and Responses APIs can be used independently depending on the use case. But, it is also productive to treat the APIs as complementary. It is not currently supported, but it is planned for the LLS Agents API to alternatively use the Responses API as its backend instead of the default Chat Completions API, i.e., enabling a combination of the safety features of Agents with the dynamic configuration and branching capabilities of Responses. + +## Feature Comparison + +| Feature | LLS Agents API | OpenAI Responses API | +|---------|------------|---------------------| +| **Conversation Management** | Linear persistent sessions | Can branch from any previous response ID | +| **Input/Output Safety Shields** | Supported | Not yet supported | +| **Per-call Flexibility** | Static per-session configuration | Dynamic per-call configuration | + +## Use Case Example: Research with Multiple Search Methods + +Let's compare how both APIs handle a research task where we need to: +1. Search for current information and examples +2. Access different information sources dynamically +3. Continue the conversation based on search results + + + + +### Session-based Configuration with Safety Shields + +```python +# Create agent with static session configuration +agent = Agent( + client, + model="Llama3.2-3B-Instruct", + instructions="You are a helpful coding assistant", + tools=[ + { + "name": "builtin::rag/knowledge_search", + "args": {"vector_db_ids": ["code_docs"]}, + }, + "builtin::code_interpreter", + ], + input_shields=["llama_guard"], + output_shields=["llama_guard"], +) + +session_id = agent.create_session("code_session") + +# First turn: Search and execute +response1 = agent.create_turn( + messages=[ + { + "role": "user", + "content": "Find examples of sorting algorithms and run a bubble sort on [3,1,4,1,5]", + }, + ], + session_id=session_id, +) + +# Continue conversation in same session +response2 = agent.create_turn( + messages=[ + { + "role": "user", + "content": "Now optimize that code and test it with a larger dataset", + }, + ], + session_id=session_id, # Same session, maintains full context +) + +# Agents API benefits: +# โœ… Safety shields protect against malicious code execution +# โœ… Session maintains context between code executions +# โœ… Consistent tool configuration throughout conversation +print(f"First result: {response1.output_message.content}") +print(f"Optimization: {response2.output_message.content}") +``` + + + + +### Dynamic Per-call Configuration with Branching + +```python +# First response: Use web search for latest algorithms +response1 = client.responses.create( + model="Llama3.2-3B-Instruct", + input="Search for the latest efficient sorting algorithms and their performance comparisons", + tools=[ + { + "type": "web_search", + }, + ], # Web search for current information +) + +# Continue conversation: Switch to file search for local docs +response2 = client.responses.create( + model="Llama3.2-1B-Instruct", # Switch to faster model + input="Now search my uploaded files for existing sorting implementations", + tools=[ + { # Using Responses API built-in tools + "type": "file_search", + "vector_store_ids": ["vs_abc123"], # Vector store containing uploaded files + }, + ], + previous_response_id=response1.id, +) + +# Branch from first response: Try different search approach +response3 = client.responses.create( + model="Llama3.2-3B-Instruct", + input="Instead, search the web for Python-specific sorting best practices", + tools=[{"type": "web_search"}], # Different web search query + previous_response_id=response1.id, # Branch from response1 +) + +# Responses API benefits: +# โœ… Dynamic tool switching (web search โ†” file search per call) +# โœ… OpenAI-compatible tool patterns (web_search, file_search) +# โœ… Branch conversations to explore different information sources +# โœ… Model flexibility per search type +print(f"Web search results: {response1.output_message.content}") +print(f"File search results: {response2.output_message.content}") +print(f"Alternative web search: {response3.output_message.content}") +``` + + + + +Both APIs demonstrate distinct strengths that make them valuable on their own for different scenarios. The Agents API excels in providing structured, safety-conscious workflows with persistent session management, while the Responses API offers flexibility through dynamic configuration and OpenAI compatible tool patterns. + +## Use Case Examples + +### 1. Research and Analysis with Safety Controls +**Best Choice: Agents API** + +**Scenario:** You're building a research assistant for a financial institution that needs to analyze market data, execute code to process financial models, and search through internal compliance documents. The system must ensure all interactions are logged for regulatory compliance and protected by safety shields to prevent malicious code execution or data leaks. + +**Why Agents API?** The Agents API provides persistent session management for iterative research workflows, built-in safety shields to protect against malicious code in financial models, and structured execution logs (session/turn/step) required for regulatory compliance. The static tool configuration ensures consistent access to your knowledge base and code interpreter throughout the entire research session. + +### 2. Dynamic Information Gathering with Branching Exploration +**Best Choice: Responses API** + +**Scenario:** You're building a competitive intelligence tool that helps businesses research market trends. Users need to dynamically switch between web search for current market data and file search through uploaded industry reports. They also want to branch conversations to explore different market segments simultaneously and experiment with different models for various analysis types. + +**Why Responses API?** The Responses API's branching capability lets users explore multiple market segments from any research point. Dynamic per-call configuration allows switching between web search and file search as needed, while experimenting with different models (faster models for quick searches, more powerful models for deep analysis). The OpenAI-compatible tool patterns make integration straightforward. + +### 3. OpenAI Migration with Advanced Tool Capabilities +**Best Choice: Responses API** + +**Scenario:** You have an existing application built with OpenAI's Assistants API that uses file search and web search capabilities. You want to migrate to Llama Stack for better performance and cost control while maintaining the same tool calling patterns and adding new capabilities like dynamic vector store selection. + +**Why Responses API?** The Responses API provides full OpenAI tool compatibility (`web_search`, `file_search`) with identical syntax, making migration seamless. The dynamic per-call configuration enables advanced features like switching vector stores per query or changing models based on query complexity - capabilities that extend beyond basic OpenAI functionality while maintaining compatibility. + +### 4. Educational Programming Tutor +**Best Choice: Agents API** + +**Scenario:** You're building a programming tutor that maintains student context across multiple sessions, safely executes code exercises, and tracks learning progress with audit trails for educators. + +**Why Agents API?** Persistent sessions remember student progress across multiple interactions, safety shields prevent malicious code execution while allowing legitimate programming exercises, and structured execution logs help educators track learning patterns. + +### 5. Advanced Software Debugging Assistant +**Best Choice: Agents API with Responses Backend** + +**Scenario:** You're building a debugging assistant that helps developers troubleshoot complex issues. It needs to maintain context throughout a debugging session, safely execute diagnostic code, switch between different analysis tools dynamically, and branch conversations to explore multiple potential causes simultaneously. + +**Why Agents + Responses?** The Agent provides safety shields for code execution and session management for the overall debugging workflow. The underlying Responses API enables dynamic model selection and flexible tool configuration per query, while branching lets you explore different theories (memory leak vs. concurrency issue) from the same debugging point and compare results. + +:::info[Future Enhancement] +The ability to use Responses API as the backend for Agents is not yet implemented but is planned for a future release. Currently, Agents use Chat Completions API as their backend by default. +::: + +## Decision Framework + +Use this framework to choose the right API for your use case: + +### Choose Agents API when: +- โœ… You need **safety shields** for input/output validation +- โœ… Your application requires **linear conversation flow** with persistent context +- โœ… You need **audit trails** and structured execution logs +- โœ… Your tool configuration is **static** throughout the session +- โœ… You're building **educational, financial, or enterprise** applications with compliance requirements + +### Choose Responses API when: +- โœ… You need **conversation branching** to explore multiple paths +- โœ… You want **dynamic per-call configuration** (models, tools, vector stores) +- โœ… You're **migrating from OpenAI** and want familiar tool patterns +- โœ… You need **OpenAI compatibility** for existing workflows +- โœ… Your application benefits from **flexible, experimental** interactions + +## Related Resources + +- **[Agents](./agent)** - Understanding the Agents API fundamentals +- **[Agent Execution Loop](./agent_execution_loop)** - How agents process turns and steps +- **[Tools Integration](./tools)** - Adding capabilities to both APIs +- **[OpenAI Compatibility](../providers/openai)** - Using OpenAI-compatible endpoints +- **[Safety Guardrails](./safety)** - Implementing safety measures in agents diff --git a/versioned_docs/version-v0.2.23/building_applications/safety.mdx b/versioned_docs/version-v0.2.23/building_applications/safety.mdx new file mode 100644 index 0000000..16fe5f6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/safety.mdx @@ -0,0 +1,395 @@ +--- +title: Safety Guardrails +description: Implement safety measures and content moderation in Llama Stack applications +sidebar_label: Safety +sidebar_position: 9 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Safety Guardrails + +Safety is a critical component of any AI application. Llama Stack provides a comprehensive Shield system that can be applied at multiple touchpoints to ensure responsible AI behavior and content moderation. + +## Shield System Overview + +The Shield system in Llama Stack provides: +- **Content filtering** for both input and output messages +- **Multi-touchpoint protection** across your application flow +- **Configurable safety policies** tailored to your use case +- **Integration with agents** for automated safety enforcement + +## Basic Shield Usage + +### Registering a Safety Shield + + + + +```python +# Register a safety shield +shield_id = "content_safety" +client.shields.register( + shield_id=shield_id, + provider_shield_id="llama-guard-basic" +) +``` + + + + +```python +# Run content through shield manually +response = client.safety.run_shield( + shield_id=shield_id, + messages=[{"role": "user", "content": "User message here"}] +) + +if response.violation: + print(f"Safety violation detected: {response.violation.user_message}") + # Handle violation appropriately +else: + print("Content passed safety checks") +``` + + + + +## Agent Integration + +Shields can be automatically applied to agent interactions for seamless safety enforcement: + + + + +```python +from llama_stack_client import Agent + +# Create agent with input safety shields +agent = Agent( + client, + model="meta-llama/Llama-3.2-3B-Instruct", + instructions="You are a helpful assistant", + input_shields=["content_safety"], # Shield user inputs + tools=["builtin::websearch"], +) + +session_id = agent.create_session("safe_session") + +# All user inputs will be automatically screened +response = agent.create_turn( + messages=[{"role": "user", "content": "Tell me about AI safety"}], + session_id=session_id, +) +``` + + + + +```python +# Create agent with output safety shields +agent = Agent( + client, + model="meta-llama/Llama-3.2-3B-Instruct", + instructions="You are a helpful assistant", + output_shields=["content_safety"], # Shield agent outputs + tools=["builtin::websearch"], +) + +session_id = agent.create_session("safe_session") + +# All agent responses will be automatically screened +response = agent.create_turn( + messages=[{"role": "user", "content": "Help me with my research"}], + session_id=session_id, +) +``` + + + + +```python +# Create agent with comprehensive safety coverage +agent = Agent( + client, + model="meta-llama/Llama-3.2-3B-Instruct", + instructions="You are a helpful assistant", + input_shields=["content_safety"], # Screen user inputs + output_shields=["content_safety"], # Screen agent outputs + tools=["builtin::websearch"], +) + +session_id = agent.create_session("fully_protected_session") + +# Both input and output are automatically protected +response = agent.create_turn( + messages=[{"role": "user", "content": "Research question here"}], + session_id=session_id, +) +``` + + + + +## Available Shield Types + +### Llama Guard Shields + +Llama Guard provides state-of-the-art content safety classification: + + + + +```python +# Basic Llama Guard for general content safety +client.shields.register( + shield_id="llama_guard_basic", + provider_shield_id="llama-guard-basic" +) +``` + +**Use Cases:** +- General content moderation +- Harmful content detection +- Basic safety compliance + + + + +```python +# Advanced Llama Guard with custom categories +client.shields.register( + shield_id="llama_guard_advanced", + provider_shield_id="llama-guard-advanced", + config={ + "categories": [ + "violence", "hate_speech", "sexual_content", + "self_harm", "illegal_activity" + ], + "threshold": 0.8 + } +) +``` + +**Use Cases:** +- Fine-tuned safety policies +- Domain-specific content filtering +- Enterprise compliance requirements + + + + +### Custom Safety Shields + +Create domain-specific safety shields for specialized use cases: + +```python +# Register custom safety shield +client.shields.register( + shield_id="financial_compliance", + provider_shield_id="custom-financial-shield", + config={ + "detect_pii": True, + "financial_advice_warning": True, + "regulatory_compliance": "FINRA" + } +) +``` + +## Safety Response Handling + +When safety violations are detected, handle them appropriately: + + + + +```python +response = client.safety.run_shield( + shield_id="content_safety", + messages=[{"role": "user", "content": "Potentially harmful content"}] +) + +if response.violation: + violation = response.violation + print(f"Violation Type: {violation.violation_type}") + print(f"User Message: {violation.user_message}") + print(f"Metadata: {violation.metadata}") + + # Log the violation for audit purposes + logger.warning(f"Safety violation detected: {violation.violation_type}") + + # Provide appropriate user feedback + return "I can't help with that request. Please try asking something else." +``` + + + + +```python +def handle_safety_response(safety_response, user_message): + """Advanced safety response handling with logging and user feedback""" + + if not safety_response.violation: + return {"safe": True, "message": "Content passed safety checks"} + + violation = safety_response.violation + + # Log violation details + audit_log = { + "timestamp": datetime.now().isoformat(), + "violation_type": violation.violation_type, + "original_message": user_message, + "shield_response": violation.user_message, + "metadata": violation.metadata + } + logger.warning(f"Safety violation: {audit_log}") + + # Determine appropriate response based on violation type + if violation.violation_type == "hate_speech": + user_feedback = "I can't engage with content that contains hate speech. Let's keep our conversation respectful." + elif violation.violation_type == "violence": + user_feedback = "I can't provide information that could promote violence. How else can I help you today?" + else: + user_feedback = "I can't help with that request. Please try asking something else." + + return { + "safe": False, + "user_feedback": user_feedback, + "violation_details": audit_log + } + +# Usage +safety_result = handle_safety_response(response, user_input) +if not safety_result["safe"]: + return safety_result["user_feedback"] +``` + + + + +## Safety Configuration Best Practices + +### ๐Ÿ›ก๏ธ **Multi-Layer Protection** +- Use both input and output shields for comprehensive coverage +- Combine multiple shield types for different threat categories +- Implement fallback mechanisms when shields fail + +### ๐Ÿ“Š **Monitoring & Auditing** +- Log all safety violations for compliance and analysis +- Monitor false positive rates to tune shield sensitivity +- Track safety metrics across different use cases + +### โš™๏ธ **Configuration Management** +- Use environment-specific safety configurations +- Implement A/B testing for shield effectiveness +- Regularly update shield models and policies + +### ๐Ÿ”ง **Integration Patterns** +- Integrate shields early in the development process +- Test safety measures with adversarial inputs +- Provide clear user feedback for violations + +## Advanced Safety Scenarios + +### Context-Aware Safety + +```python +# Safety shields that consider conversation context +agent = Agent( + client, + model="meta-llama/Llama-3.2-3B-Instruct", + instructions="You are a healthcare assistant", + input_shields=["medical_safety"], + output_shields=["medical_safety"], + # Context helps shields make better decisions + safety_context={ + "domain": "healthcare", + "user_type": "patient", + "compliance_level": "HIPAA" + } +) +``` + +### Dynamic Shield Selection + +```python +def select_shield_for_user(user_profile): + """Select appropriate safety shield based on user context""" + if user_profile.age < 18: + return "child_safety_shield" + elif user_profile.context == "enterprise": + return "enterprise_compliance_shield" + else: + return "general_safety_shield" + +# Use dynamic shield selection +shield_id = select_shield_for_user(current_user) +response = client.safety.run_shield( + shield_id=shield_id, + messages=messages +) +``` + +## Compliance and Regulations + +### Industry-Specific Safety + + + + +```python +# Healthcare-specific safety configuration +client.shields.register( + shield_id="hipaa_compliance", + provider_shield_id="healthcare-safety-shield", + config={ + "detect_phi": True, # Protected Health Information + "medical_advice_warning": True, + "regulatory_framework": "HIPAA" + } +) +``` + + + + +```python +# Financial services safety configuration +client.shields.register( + shield_id="finra_compliance", + provider_shield_id="financial-safety-shield", + config={ + "detect_financial_advice": True, + "investment_disclaimers": True, + "regulatory_framework": "FINRA" + } +) +``` + + + + +```python +# Educational platform safety for minors +client.shields.register( + shield_id="coppa_compliance", + provider_shield_id="educational-safety-shield", + config={ + "child_protection": True, + "educational_content_only": True, + "regulatory_framework": "COPPA" + } +) +``` + + + + +## Related Resources + +- **[Agents](./agent)** - Integrating safety shields with intelligent agents +- **[Agent Execution Loop](./agent_execution_loop)** - Understanding safety in the execution flow +- **[Evaluations](./evals)** - Evaluating safety shield effectiveness +- **[Telemetry](./telemetry)** - Monitoring safety violations and metrics +- **[Llama Guard Documentation](https://github.com/meta-llama/PurpleLlama/tree/main/Llama-Guard3)** - Advanced safety model details diff --git a/versioned_docs/version-v0.2.23/building_applications/telemetry.mdx b/versioned_docs/version-v0.2.23/building_applications/telemetry.mdx new file mode 100644 index 0000000..6a255e7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/telemetry.mdx @@ -0,0 +1,342 @@ +--- +title: Telemetry +description: Monitor and observe Llama Stack applications with comprehensive telemetry capabilities +sidebar_label: Telemetry +sidebar_position: 8 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Telemetry + +The Llama Stack telemetry system provides comprehensive tracing, metrics, and logging capabilities. It supports multiple sink types including OpenTelemetry, SQLite, and Console output for complete observability of your AI applications. + +## Event Types + +The telemetry system supports three main types of events: + + + + +Free-form log messages with severity levels for general application logging: + +```python +unstructured_log_event = UnstructuredLogEvent( + message="This is a log message", + severity=LogSeverity.INFO +) +``` + + + + +Numerical measurements with units for tracking performance and usage: + +```python +metric_event = MetricEvent( + metric="my_metric", + value=10, + unit="count" +) +``` + + + + +System events like span start/end that provide structured operation tracking: + +```python +structured_log_event = SpanStartPayload( + name="my_span", + parent_span_id="parent_span_id" +) +``` + + + + +## Spans and Traces + +- **Spans**: Represent individual operations with timing information and hierarchical relationships +- **Traces**: Collections of related spans that form a complete request flow across your application + +This hierarchical structure allows you to understand the complete execution path of requests through your Llama Stack application. + +## Automatic Metrics Generation + +Llama Stack automatically generates metrics during inference operations. These metrics are aggregated at the **inference request level** and provide insights into token usage and model performance. + +### Available Metrics + +The following metrics are automatically generated for each inference request: + +| Metric Name | Type | Unit | Description | Labels | +|-------------|------|------|-------------|--------| +| `llama_stack_prompt_tokens_total` | Counter | `tokens` | Number of tokens in the input prompt | `model_id`, `provider_id` | +| `llama_stack_completion_tokens_total` | Counter | `tokens` | Number of tokens in the generated response | `model_id`, `provider_id` | +| `llama_stack_tokens_total` | Counter | `tokens` | Total tokens used (prompt + completion) | `model_id`, `provider_id` | + +### Metric Generation Flow + +1. **Token Counting**: During inference operations (chat completion, completion, etc.), the system counts tokens in both input prompts and generated responses +2. **Metric Construction**: For each request, `MetricEvent` objects are created with the token counts +3. **Telemetry Logging**: Metrics are sent to the configured telemetry sinks +4. **OpenTelemetry Export**: When OpenTelemetry is enabled, metrics are exposed as standard OpenTelemetry counters + +### Metric Aggregation Level + +All metrics are generated and aggregated at the **inference request level**. This means: + +- Each individual inference request generates its own set of metrics +- Metrics are not pre-aggregated across multiple requests +- Aggregation (sums, averages, etc.) can be performed by your observability tools (Prometheus, Grafana, etc.) +- Each metric includes labels for `model_id` and `provider_id` to enable filtering and grouping + +### Example Metric Event + +```python +MetricEvent( + trace_id="1234567890abcdef", + span_id="abcdef1234567890", + metric="total_tokens", + value=150, + timestamp=1703123456.789, + unit="tokens", + attributes={ + "model_id": "meta-llama/Llama-3.2-3B-Instruct", + "provider_id": "tgi" + }, +) +``` + +## Telemetry Sinks + +Choose from multiple sink types based on your observability needs: + + + + +Send events to an OpenTelemetry Collector for integration with observability platforms: + +**Use Cases:** +- Visualizing traces in tools like Jaeger +- Collecting metrics for Prometheus +- Integration with enterprise observability stacks + +**Features:** +- Standard OpenTelemetry format +- Compatible with all OpenTelemetry collectors +- Supports both traces and metrics + + + + +Store events in a local SQLite database for direct querying: + +**Use Cases:** +- Local development and debugging +- Custom analytics and reporting +- Offline analysis of application behavior + +**Features:** +- Direct SQL querying capabilities +- Persistent local storage +- No external dependencies + + + + +Print events to the console for immediate debugging: + +**Use Cases:** +- Development and testing +- Quick debugging sessions +- Simple logging without external tools + +**Features:** +- Immediate output visibility +- No setup required +- Human-readable format + + + + +## Configuration + +### Meta-Reference Provider + +Currently, only the meta-reference provider is implemented. It can be configured to send events to multiple sink types: + +```yaml +telemetry: + - provider_id: meta-reference + provider_type: inline::meta-reference + config: + service_name: "llama-stack-service" + sinks: ['console', 'sqlite', 'otel_trace', 'otel_metric'] + otel_exporter_otlp_endpoint: "http://localhost:4318" + sqlite_db_path: "/path/to/telemetry.db" +``` + +### Environment Variables + +Configure telemetry behavior using environment variables: + +- **`OTEL_EXPORTER_OTLP_ENDPOINT`**: OpenTelemetry Collector endpoint (default: `http://localhost:4318`) +- **`OTEL_SERVICE_NAME`**: Service name for telemetry (default: empty string) +- **`TELEMETRY_SINKS`**: Comma-separated list of sinks (default: `console,sqlite`) + +## Visualization with Jaeger + +The `otel_trace` sink works with any service compatible with the OpenTelemetry collector. Traces and metrics use separate endpoints but can share the same collector. + +### Starting Jaeger + +Start a Jaeger instance with OTLP HTTP endpoint at 4318 and the Jaeger UI at 16686: + +```bash +docker run --pull always --rm --name jaeger \ + -p 16686:16686 -p 4318:4318 \ + jaegertracing/jaeger:2.1.0 +``` + +Once running, you can visualize traces by navigating to [http://localhost:16686/](http://localhost:16686/). + +## Querying Metrics + +When using the OpenTelemetry sink, metrics are exposed in standard format and can be queried through various tools: + + + + +Example Prometheus queries for analyzing token usage: + +```promql +# Total tokens used across all models +sum(llama_stack_tokens_total) + +# Tokens per model +sum by (model_id) (llama_stack_tokens_total) + +# Average tokens per request over 5 minutes +rate(llama_stack_tokens_total[5m]) + +# Token usage by provider +sum by (provider_id) (llama_stack_tokens_total) +``` + + + + +Create dashboards using Prometheus as a data source: + +- **Token Usage Over Time**: Line charts showing token consumption trends +- **Model Performance**: Comparison of different models by token efficiency +- **Provider Analysis**: Breakdown of usage across different providers +- **Request Patterns**: Understanding peak usage times and patterns + + + + +Forward metrics to other observability systems: + +- Export to multiple backends simultaneously +- Apply transformations and filtering +- Integrate with existing monitoring infrastructure + + + + +## SQLite Querying + +The `sqlite` sink allows you to query traces without an external system. This is particularly useful for development and custom analytics. + +### Example Queries + +```sql +-- Query recent traces +SELECT * FROM traces WHERE timestamp > datetime('now', '-1 hour'); + +-- Analyze span durations +SELECT name, AVG(duration_ms) as avg_duration +FROM spans +GROUP BY name +ORDER BY avg_duration DESC; + +-- Find slow operations +SELECT * FROM spans +WHERE duration_ms > 1000 +ORDER BY duration_ms DESC; +``` + +:::tip[Advanced Analytics] +Refer to the [Getting Started notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb) for more examples on querying traces and spans programmatically. +::: + +## Best Practices + +### ๐Ÿ” **Monitoring Strategy** +- Use OpenTelemetry for production environments +- Combine multiple sinks for development (console + SQLite) +- Set up alerts on key metrics like token usage and error rates + +### ๐Ÿ“Š **Metrics Analysis** +- Track token usage trends to optimize costs +- Monitor response times across different models +- Analyze usage patterns to improve resource allocation + +### ๐Ÿšจ **Alerting & Debugging** +- Set up alerts for unusual token consumption spikes +- Use trace data to debug performance issues +- Monitor error rates and failure patterns + +### ๐Ÿ”ง **Configuration Management** +- Use environment variables for flexible deployment +- Configure appropriate retention policies for SQLite +- Ensure proper network access to OpenTelemetry collectors + +## Integration Examples + +### Basic Telemetry Setup + +```python +from llama_stack_client import LlamaStackClient + +# Client with telemetry headers +client = LlamaStackClient( + base_url="http://localhost:8000", + extra_headers={ + "X-Telemetry-Service": "my-ai-app", + "X-Telemetry-Version": "1.0.0" + } +) + +# All API calls will be automatically traced +response = client.inference.chat_completion( + model="meta-llama/Llama-3.2-3B-Instruct", + messages=[{"role": "user", "content": "Hello!"}] +) +``` + +### Custom Telemetry Context + +```python +# Add custom span attributes for better tracking +with tracer.start_as_current_span("custom_operation") as span: + span.set_attribute("user_id", "user123") + span.set_attribute("operation_type", "chat_completion") + + response = client.inference.chat_completion( + model="meta-llama/Llama-3.2-3B-Instruct", + messages=[{"role": "user", "content": "Hello!"}] + ) +``` + +## Related Resources + +- **[Agents](./agent)** - Monitoring agent execution with telemetry +- **[Evaluations](./evals)** - Using telemetry data for performance evaluation +- **[Getting Started Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb)** - Telemetry examples and queries +- **[OpenTelemetry Documentation](https://opentelemetry.io/)** - Comprehensive observability framework +- **[Jaeger Documentation](https://www.jaegertracing.io/)** - Distributed tracing visualization diff --git a/versioned_docs/version-v0.2.23/building_applications/tools.mdx b/versioned_docs/version-v0.2.23/building_applications/tools.mdx new file mode 100644 index 0000000..be60a16 --- /dev/null +++ b/versioned_docs/version-v0.2.23/building_applications/tools.mdx @@ -0,0 +1,340 @@ +--- +title: Tools +description: Extend agent capabilities with external tools and function calling +sidebar_label: Tools +sidebar_position: 6 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Tools + +Tools are functions that can be invoked by an agent to perform tasks. They are organized into tool groups and registered with specific providers. Each tool group represents a collection of related tools from a single provider. They are organized into groups so that state can be externalized: the collection operates on the same state typically. + +An example of this would be a "db_access" tool group that contains tools for interacting with a database. "list_tables", "query_table", "insert_row" could be examples of tools in this group. + +Tools are treated as any other resource in llama stack like models. You can register them, have providers for them etc. + +When instantiating an agent, you can provide it a list of tool groups that it has access to. Agent gets the corresponding tool definitions for the specified tool groups and passes them along to the model. + +Refer to the [Building AI Applications](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb) notebook for more examples on how to use tools. + +## Server-side vs. Client-side Tool Execution + +Llama Stack allows you to use both server-side and client-side tools. With server-side tools, `agent.create_turn` can perform execution of the tool calls emitted by the model transparently giving the user the final answer desired. If client-side tools are provided, the tool call is sent back to the user for execution and optional continuation using the `agent.resume_turn` method. + +## Server-side Tools + +Llama Stack provides built-in providers for some common tools. These include web search, math, and RAG capabilities. + +### Web Search + +You have three providers to execute the web search tool calls generated by a model: Brave Search, Bing Search, and Tavily Search. + +To indicate that the web search tool calls should be executed by brave-search, you can point the "builtin::websearch" toolgroup to the "brave-search" provider. + +```python +client.toolgroups.register( + toolgroup_id="builtin::websearch", + provider_id="brave-search", + args={"max_results": 5}, +) +``` + +The tool requires an API key which can be provided either in the configuration or through the request header `X-LlamaStack-Provider-Data`. The format of the header is: +``` +{"_api_key": } +``` + +### Math + +The WolframAlpha tool provides access to computational knowledge through the WolframAlpha API. + +```python +client.toolgroups.register( + toolgroup_id="builtin::wolfram_alpha", + provider_id="wolfram-alpha" +) +``` + +Example usage: +```python +result = client.tool_runtime.invoke_tool( + tool_name="wolfram_alpha", + args={"query": "solve x^2 + 2x + 1 = 0"} +) +``` + +### RAG + +The RAG tool enables retrieval of context from various types of memory banks (vector, key-value, keyword, and graph). + +```python +# Register Memory tool group +client.toolgroups.register( + toolgroup_id="builtin::rag", + provider_id="faiss", + args={"max_chunks": 5, "max_tokens_in_context": 4096}, +) +``` + +Features: +- Support for multiple memory bank types +- Configurable query generation +- Context retrieval with token limits + +:::note[Default Configuration] +By default, llama stack run.yaml defines toolgroups for web search, wolfram alpha and rag, that are provided by tavily-search, wolfram-alpha and rag providers. +::: + +## Model Context Protocol (MCP) + +[MCP](https://github.com/modelcontextprotocol) is an upcoming, popular standard for tool discovery and execution. It is a protocol that allows tools to be dynamically discovered from an MCP endpoint and can be used to extend the agent's capabilities. + +### Using Remote MCP Servers + +You can find some popular remote MCP servers [here](https://github.com/jaw9c/awesome-remote-mcp-servers). You can register them as toolgroups in the same way as local providers. + +```python +client.toolgroups.register( + toolgroup_id="mcp::deepwiki", + provider_id="model-context-protocol", + mcp_endpoint=URL(uri="https://mcp.deepwiki.com/sse"), +) +``` + +Note that most of the more useful MCP servers need you to authenticate with them. Many of them use OAuth2.0 for authentication. You can provide authorization headers to send to the MCP server using the "Provider Data" abstraction provided by Llama Stack. When making an agent call, + +```python +agent = Agent( + ..., + tools=["mcp::deepwiki"], + extra_headers={ + "X-LlamaStack-Provider-Data": json.dumps( + { + "mcp_headers": { + "http://mcp.deepwiki.com/sse": { + "Authorization": "Bearer ", + }, + }, + } + ), + }, +) +agent.create_turn(...) +``` + +### Running Your Own MCP Server + +Here's an example of how to run a simple MCP server that exposes a File System as a set of tools to the Llama Stack agent. + + + + +```shell +# Start your MCP server +mkdir /tmp/content +touch /tmp/content/foo +touch /tmp/content/bar +npx -y supergateway --port 8000 --stdio 'npx -y @modelcontextprotocol/server-filesystem /tmp/content' +``` + + + + +```python +# Register the MCP server as a tool group +client.toolgroups.register( + toolgroup_id="mcp::filesystem", + provider_id="model-context-protocol", + mcp_endpoint=URL(uri="http://localhost:8000/sse"), +) +``` + + + + +## Adding Custom (Client-side) Tools + +When you want to use tools other than the built-in tools, you just need to implement a python function with a docstring. The content of the docstring will be used to describe the tool and the parameters and passed along to the generative model. + +```python +# Example tool definition +def my_tool(input: int) -> int: + """ + Runs my awesome tool. + + :param input: some int parameter + """ + return input * 2 +``` + +:::tip[Documentation Best Practices] +We employ python docstrings to describe the tool and the parameters. It is important to document the tool and the parameters so that the model can use the tool correctly. It is recommended to experiment with different docstrings to see how they affect the model's behavior. +::: + +Once defined, simply pass the tool to the agent config. `Agent` will take care of the rest (calling the model with the tool definition, executing the tool, and returning the result to the model for the next iteration). + +```python +# Example agent config with client provided tools +agent = Agent(client, ..., tools=[my_tool]) +``` + +Refer to [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/blob/main/examples/agents/e2e_loop_with_client_tools.py) for an example of how to use client provided tools. + +## Tool Invocation + +Tools can be invoked using the `invoke_tool` method: + +```python +result = client.tool_runtime.invoke_tool( + tool_name="web_search", + kwargs={"query": "What is the capital of France?"} +) +``` + +The result contains: +- `content`: The tool's output +- `error_message`: Optional error message if the tool failed +- `error_code`: Optional error code if the tool failed + +## Listing Available Tools + +You can list all available tools or filter by tool group: + +```python +# List all tools +all_tools = client.tools.list_tools() + +# List tools in a specific group +group_tools = client.tools.list_tools(toolgroup_id="search_tools") +``` + +## Complete Examples + +### Web Search Agent + + + + +1. Start by registering a Tavily API key at [Tavily](https://tavily.com/). +2. [Optional] Provide the API key directly to the Llama Stack server +```bash +export TAVILY_SEARCH_API_KEY="your key" +``` +```bash +--env TAVILY_SEARCH_API_KEY=${TAVILY_SEARCH_API_KEY} +``` + + + + +```python +from llama_stack_client.lib.agents.agent import Agent +from llama_stack_client.types.agent_create_params import AgentConfig +from llama_stack_client.lib.agents.event_logger import EventLogger +from llama_stack_client import LlamaStackClient + +client = LlamaStackClient( + base_url=f"http://localhost:8321", + provider_data={ + "tavily_search_api_key": "your_TAVILY_SEARCH_API_KEY" + }, # Set this from the client side. No need to provide it if it has already been configured on the Llama Stack server. +) + +agent = Agent( + client, + model="meta-llama/Llama-3.2-3B-Instruct", + instructions=( + "You are a web search assistant, must use websearch tool to look up the most current and precise information available. " + ), + tools=["builtin::websearch"], +) + +session_id = agent.create_session("websearch-session") + +response = agent.create_turn( + messages=[ + {"role": "user", "content": "How did the USA perform in the last Olympics?"} + ], + session_id=session_id, +) +for log in EventLogger().log(response): + log.print() +``` + + + + +### WolframAlpha Math Agent + + + + +1. Start by registering for a WolframAlpha API key at [WolframAlpha Developer Portal](https://developer.wolframalpha.com/access). +2. Provide the API key either when starting the Llama Stack server: + ```bash + --env WOLFRAM_ALPHA_API_KEY=${WOLFRAM_ALPHA_API_KEY} + ``` + or from the client side: + ```python + client = LlamaStackClient( + base_url="http://localhost:8321", + provider_data={"wolfram_alpha_api_key": wolfram_api_key}, + ) + ``` + + + + +```python +# Configure the tools in the Agent by setting tools=["builtin::wolfram_alpha"] +agent = Agent( + client, + model="meta-llama/Llama-3.2-3B-Instruct", + instructions="You are a mathematical assistant that can solve complex equations.", + tools=["builtin::wolfram_alpha"], +) + +session_id = agent.create_session("math-session") + +# Example user query +response = agent.create_turn( + messages=[{"role": "user", "content": "Solve x^2 + 2x + 1 = 0 using WolframAlpha"}], + session_id=session_id, +) +``` + + + + +## Best Practices + +### ๐Ÿ› ๏ธ **Tool Selection** +- Use **server-side tools** for production applications requiring reliability and security +- Use **client-side tools** for development, prototyping, or specialized integrations +- Combine multiple tool types for comprehensive functionality + +### ๐Ÿ“ **Documentation** +- Write clear, detailed docstrings for custom tools +- Include parameter descriptions and expected return types +- Test tool descriptions with the model to ensure proper usage + +### ๐Ÿ” **Security** +- Store API keys securely using environment variables or secure configuration +- Use the `X-LlamaStack-Provider-Data` header for dynamic authentication +- Validate tool inputs and outputs for security + +### ๐Ÿ”„ **Error Handling** +- Implement proper error handling in custom tools +- Use structured error responses with meaningful messages +- Monitor tool performance and reliability + +## Related Resources + +- **[Agents](./agent)** - Building intelligent agents with tools +- **[RAG (Retrieval Augmented Generation)](./rag)** - Using knowledge retrieval tools +- **[Agent Execution Loop](./agent_execution_loop)** - Understanding tool execution flow +- **[Building AI Applications Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb)** - Comprehensive examples +- **[Llama Stack Apps Examples](https://github.com/meta-llama/llama-stack-apps)** - Real-world tool implementations diff --git a/versioned_docs/version-v0.2.23/concepts/apis/api_leveling.mdx b/versioned_docs/version-v0.2.23/concepts/apis/api_leveling.mdx new file mode 100644 index 0000000..e3e118d --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/apis/api_leveling.mdx @@ -0,0 +1,101 @@ +--- +title: API Stability Leveling +description: Understanding API stability levels and versioning in Llama Stack +sidebar_label: API Stability +sidebar_position: 4 +--- + +# Llama Stack API Stability Leveling + +In order to provide a stable experience in Llama Stack, the various APIs need different stability levels indicating the level of support, backwards compatability, and overall production readiness. + +## Different Levels + +### v1alpha + +- Little to no expectation of support between versions +- Breaking changes are permitted +- Datatypes and parameters can break +- Routes can be added and removed + +#### Graduation Criteria + +- an API can graduate from `v1alpha` to `v1beta` if the team has identified the extent of the non-optional routes and the shape of their parameters/return types for the API eg. `/v1/openai/chat/completions`. Optional types can change. +- CRUD must stay stable once in `v1beta`. This is a commitment to backward compatibility, guaranteeing that most code you write against the v1beta version will not break during future updates. We may make additive changes (like adding a new, optional field to a response), but we will not make breaking changes (like renaming an existing "modelName" field to "name", changing an ID's data type from an integer to a string, or altering an endpoint URL). +- for OpenAI APIs, a comparison to the OpenAI spec for the specific API can be done to ensure completeness. + +### v1beta + +- API routes remain consistent between versions +- Parameters and return types are not ensured between versions +- API, besides minor fixes and adjustments, should be _almost_ v1. Changes should not be drastic. + +#### Graduation Criteria + +- an API can graduate from `v1beta` to `v1` if the API surface and datatypes are complete as identified by the team. The parameters and return types that are mandatory for each route are stable. All aspects of graduating from `v1alpha1` to `v1beta` apply as well. +- Optional parameters, routes, or parts of the return type can be added after graduating to `v1` + +### v1 (stable) + +- Considered stable +- Backwards compatible between Z-streams + - Y-stream breaking changes must go through the proper approval and announcement process. +- Datatypes for a route and its return types cannot change between Z-streams + - Y-stream datatype changes should be sparing, unless the changes are additional net-new parameters +- Must have proper conformance testing as outlined in https://github.com/llamastack/llama-stack/issues/3237 + +### v2+ (Major Versions) + +Introducing a new major version like `/v2` is a significant and disruptive event that should be treated as a last resort. It is reserved for essential changes to a stable `/v1` API that are fundamentally backward-incompatible and cannot be implemented through additive, non-breaking changes or breaking changes across X/Y-Stream releases (x.y.z). + +If a `/v2` version is deemed absolutely necessary, it must adhere to the following protocol to ensure a sane and predictable transition for users: + +#### Lifecycle Progression + + A new major version must follow the same stability lifecycle as `/v1`. It will be introduced as `/v2alpha`, mature to `/v2beta`, and finally become stable as `/v2`. + +#### Coexistence: + +The new `/v2` API must be introduced alongside the existing `/v1` API and run in parallel. It must not replace the `/v1` API immediately. + +#### Deprecation Policy: + +When a `/v2` API is introduced, a clear and generous deprecation policy for the `/v1` API must be published simultaneously. This policy must outline the timeline for the eventual removal of the `/v1` API, giving users ample time to migrate. + +### API Stability vs. Provider Stability + +The leveling introduced in this document relates to the stability of the API and not specifically the providers within the API. + +Providers can iterate as much as they want on functionality as long as they work within the bounds of an API. If they need to change the API, then the API should not be `/v1`, or those breaking changes can only happen on a y-stream release basis. + +### Approval and Announcement Process for Breaking Changes + +- **PR Labeling**: Any pull request that introduces a breaking API change must be clearly labeled with `breaking-change`. +- **PR Title/Commit**: Any pull request that introduces a breaking API change must contain `BREAKING CHANGE` in the title and commit footer. Alternatively, the commit can include `!`, eg. `feat(api)!: title goes here` This is outlined in the [conventional commits documentation](https://www.conventionalcommits.org/en/v1.0.0/#specification) +- **Maintainer Review**: At least one maintainer must explicitly acknowledge the breaking change during review by applying the `breaking-change` label. An approval must come with this label or the acknowledgement this label has already been applied. +- **Announcement**: Breaking changes require inclusion in release notes and, if applicable, a separate communication (e.g., Discord, Github Issues, or GitHub Discussions) prior to release. + +If a PR has proper approvals, labels, and commit/title hygiene, the failing API conformance tests will be bypassed. + + +## Enforcement + +### Migration of API routes under `/v1alpha`, `/v1beta`, and `/v1` + +Instead of placing every API under `/v1`, any API that is not fully stable or complete should go under `/v1alpha` or `/v1beta`. For example, at the time of this writing, `post_training` belongs here, as well as any OpenAI-compatible API whose surface does not exactly match the upstream OpenAI API it mimics. + +This migration is crucial as we get Llama Stack in the hands of users who intend to productize various APIs. A clear view of what is stable and what is actively being developed will enable users to pick and choose various APIs to build their products on. + +This migration will be a breaking change for any API moving out of `/v1`. Ideally, this should happen before 0.3.0 and especially 1.0.0. + +### `x-stability` tags in the OpenAPI spec for oasdiff + +`x-stability` tags allow tools like oasdiff to enforce different rules for different stability levels; these tags should match the routes: [oasdiff stability](https://github.com/oasdiff/oasdiff/blob/main/docs/STABILITY.md) + +### Testing + +The testing of each stable API is already outlined in [issue #3237](https://github.com/llamastack/llama-stack/issues/3237) and is being worked on. These sorts of conformance tests should apply primarily to `/v1` APIs only, with `/v1alpha` and `/v1beta` having any tests the maintainers see fit as well as basic testing to ensure the routing works properly. + +### New APIs going forward + +Any subsequently introduced APIs should be introduced as `/v1alpha` diff --git a/versioned_docs/version-v0.2.23/concepts/apis/api_providers.mdx b/versioned_docs/version-v0.2.23/concepts/apis/api_providers.mdx new file mode 100644 index 0000000..5f0fe2a --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/apis/api_providers.mdx @@ -0,0 +1,19 @@ +--- +title: API Providers +description: Understanding remote vs inline provider implementations +sidebar_label: API Providers +sidebar_position: 2 +--- + +# API Providers + +The goal of Llama Stack is to build an ecosystem where users can easily swap out different implementations for the same API. Examples for these include: +- LLM inference providers (e.g., Fireworks, Together, AWS Bedrock, Groq, Cerebras, SambaNova, vLLM, etc.), +- Vector databases (e.g., ChromaDB, Weaviate, Qdrant, Milvus, FAISS, PGVector, etc.), +- Safety providers (e.g., Meta's Llama Guard, AWS Bedrock Guardrails, etc.) + +Providers come in two flavors: +- **Remote**: the provider runs as a separate service external to the Llama Stack codebase. Llama Stack contains a small amount of adapter code. +- **Inline**: the provider is fully specified and implemented within the Llama Stack codebase. It may be a simple wrapper around an existing library, or a full fledged implementation within Llama Stack. + +Most importantly, Llama Stack always strives to provide at least one fully inline provider for each API so you can iterate on a fully featured environment locally. diff --git a/versioned_docs/version-v0.2.23/concepts/apis/external.mdx b/versioned_docs/version-v0.2.23/concepts/apis/external.mdx new file mode 100644 index 0000000..7b4a3e8 --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/apis/external.mdx @@ -0,0 +1,398 @@ +--- +title: External APIs +description: Understanding external APIs in Llama Stack +sidebar_label: External APIs +sidebar_position: 3 +--- +# External APIs + +Llama Stack supports external APIs that live outside of the main codebase. This allows you to: +- Create and maintain your own APIs independently +- Share APIs with others without contributing to the main codebase +- Keep API-specific code separate from the core Llama Stack code + +## Configuration + +To enable external APIs, you need to configure the `external_apis_dir` in your Llama Stack configuration. This directory should contain your external API specifications: + +```yaml +external_apis_dir: ~/.llama/apis.d/ +``` + +## Directory Structure + +The external APIs directory should follow this structure: + +``` +apis.d/ + custom_api1.yaml + custom_api2.yaml +``` + +Each YAML file in these directories defines an API specification. + +## API Specification + +Here's an example of an external API specification for a weather API: + +```yaml +module: weather +api_dependencies: + - inference +protocol: WeatherAPI +name: weather +pip_packages: + - llama-stack-api-weather +``` + +### API Specification Fields + +- `module`: Python module containing the API implementation +- `protocol`: Name of the protocol class for the API +- `name`: Name of the API +- `pip_packages`: List of pip packages to install the API, typically a single package + +## Required Implementation + +External APIs must expose a `available_providers()` function in their module that returns a list of provider names: + +```python +# llama_stack_api_weather/api.py +from llama_stack.providers.datatypes import Api, InlineProviderSpec, ProviderSpec + + +def available_providers() -> list[ProviderSpec]: + return [ + InlineProviderSpec( + api=Api.weather, + provider_type="inline::darksky", + pip_packages=[], + module="llama_stack_provider_darksky", + config_class="llama_stack_provider_darksky.DarkSkyWeatherImplConfig", + ), + ] +``` + +A Protocol class like so: + +```python +# llama_stack_api_weather/api.py +from typing import Protocol + +from llama_stack.schema_utils import webmethod + + +class WeatherAPI(Protocol): + """ + A protocol for the Weather API. + """ + + @webmethod(route="/locations", method="GET") + async def get_available_locations() -> dict[str, list[str]]: + """ + Get the available locations. + """ + ... +``` + +## Example: Custom API + +Here's a complete example of creating and using a custom API: + +1. First, create the API package: + +```bash +mkdir -p llama-stack-api-weather +cd llama-stack-api-weather +mkdir src/llama_stack_api_weather +git init +uv init +``` + +2. Edit `pyproject.toml`: + +```toml +[project] +name = "llama-stack-api-weather" +version = "0.1.0" +description = "Weather API for Llama Stack" +readme = "README.md" +requires-python = ">=3.12" +dependencies = ["llama-stack", "pydantic"] + +[build-system] +requires = ["setuptools"] +build-backend = "setuptools.build_meta" + +[tool.setuptools.packages.find] +where = ["src"] +include = ["llama_stack_api_weather", "llama_stack_api_weather.*"] +``` + +3. Create the initial files: + +```bash +touch src/llama_stack_api_weather/__init__.py +touch src/llama_stack_api_weather/api.py +``` + +```python +# llama-stack-api-weather/src/llama_stack_api_weather/__init__.py +"""Weather API for Llama Stack.""" + +from .api import WeatherAPI, available_providers + +__all__ = ["WeatherAPI", "available_providers"] +``` + +4. Create the API implementation: + +```python +# llama-stack-api-weather/src/llama_stack_api_weather/weather.py +from typing import Protocol + +from llama_stack.providers.datatypes import ( + AdapterSpec, + Api, + ProviderSpec, + RemoteProviderSpec, +) +from llama_stack.schema_utils import webmethod + + +def available_providers() -> list[ProviderSpec]: + return [ + RemoteProviderSpec( + api=Api.weather, + provider_type="remote::kaze", + config_class="llama_stack_provider_kaze.KazeProviderConfig", + adapter=AdapterSpec( + adapter_type="kaze", + module="llama_stack_provider_kaze", + pip_packages=["llama_stack_provider_kaze"], + config_class="llama_stack_provider_kaze.KazeProviderConfig", + ), + ), + ] + + +class WeatherProvider(Protocol): + """ + A protocol for the Weather API. + """ + + @webmethod(route="/weather/locations", method="GET") + async def get_available_locations() -> dict[str, list[str]]: + """ + Get the available locations. + """ + ... +``` + +5. Create the API specification: + +```yaml +# ~/.llama/apis.d/weather.yaml +module: llama_stack_api_weather +name: weather +pip_packages: ["llama-stack-api-weather"] +protocol: WeatherProvider + +``` + +6. Install the API package: + +```bash +uv pip install -e . +``` + +7. Configure Llama Stack to use external APIs: + +```yaml +version: "2" +image_name: "llama-stack-api-weather" +apis: + - weather +providers: {} +external_apis_dir: ~/.llama/apis.d +``` + +The API will now be available at `/v1/weather/locations`. + +## Example: custom provider for the weather API + +1. Create the provider package: + +```bash +mkdir -p llama-stack-provider-kaze +cd llama-stack-provider-kaze +uv init +``` + +2. Edit `pyproject.toml`: + +```toml +[project] +name = "llama-stack-provider-kaze" +version = "0.1.0" +description = "Kaze weather provider for Llama Stack" +readme = "README.md" +requires-python = ">=3.12" +dependencies = ["llama-stack", "pydantic", "aiohttp"] + +[build-system] +requires = ["setuptools"] +build-backend = "setuptools.build_meta" + +[tool.setuptools.packages.find] +where = ["src"] +include = ["llama_stack_provider_kaze", "llama_stack_provider_kaze.*"] +``` + +3. Create the initial files: + +```bash +touch src/llama_stack_provider_kaze/__init__.py +touch src/llama_stack_provider_kaze/kaze.py +``` + +4. Create the provider implementation: + + +Initialization function: + +```python +# llama-stack-provider-kaze/src/llama_stack_provider_kaze/__init__.py +"""Kaze weather provider for Llama Stack.""" + +from .config import KazeProviderConfig +from .kaze import WeatherKazeAdapter + +__all__ = ["KazeProviderConfig", "WeatherKazeAdapter"] + + +async def get_adapter_impl(config: KazeProviderConfig, _deps): + from .kaze import WeatherKazeAdapter + + impl = WeatherKazeAdapter(config) + await impl.initialize() + return impl +``` + +Configuration: + +```python +# llama-stack-provider-kaze/src/llama_stack_provider_kaze/config.py +from pydantic import BaseModel, Field + + +class KazeProviderConfig(BaseModel): + """Configuration for the Kaze weather provider.""" + + base_url: str = Field( + "https://api.kaze.io/v1", + description="Base URL for the Kaze weather API", + ) +``` + +Main implementation: + +```python +# llama-stack-provider-kaze/src/llama_stack_provider_kaze/kaze.py +from llama_stack_api_weather.api import WeatherProvider + +from .config import KazeProviderConfig + + +class WeatherKazeAdapter(WeatherProvider): + """Kaze weather provider implementation.""" + + def __init__( + self, + config: KazeProviderConfig, + ) -> None: + self.config = config + + async def initialize(self) -> None: + pass + + async def get_available_locations(self) -> dict[str, list[str]]: + """Get available weather locations.""" + return {"locations": ["Paris", "Tokyo"]} +``` + +5. Create the provider specification: + +```yaml +# ~/.llama/providers.d/remote/weather/kaze.yaml +adapter: + adapter_type: kaze + pip_packages: ["llama_stack_provider_kaze"] + config_class: llama_stack_provider_kaze.config.KazeProviderConfig + module: llama_stack_provider_kaze +optional_api_dependencies: [] +``` + +6. Install the provider package: + +```bash +uv pip install -e . +``` + +7. Configure Llama Stack to use the provider: + +```yaml +# ~/.llama/run-byoa.yaml +version: "2" +image_name: "llama-stack-api-weather" +apis: + - weather +providers: + weather: + - provider_id: kaze + provider_type: remote::kaze + config: {} +external_apis_dir: ~/.llama/apis.d +external_providers_dir: ~/.llama/providers.d +server: + port: 8321 +``` + +8. Run the server: + +```bash +python -m llama_stack.core.server.server --yaml-config ~/.llama/run-byoa.yaml +``` + +9. Test the API: + +```bash +curl -sSf http://127.0.0.1:8321/v1/weather/locations +{"locations":["Paris","Tokyo"]}% +``` + +## Best Practices + +1. **Package Naming**: Use a clear and descriptive name for your API package. + +2. **Version Management**: Keep your API package versioned and compatible with the Llama Stack version you're using. + +3. **Dependencies**: Only include the minimum required dependencies in your API package. + +4. **Documentation**: Include clear documentation in your API package about: + - Installation requirements + - Configuration options + - API endpoints and usage + - Any limitations or known issues + +5. **Testing**: Include tests in your API package to ensure it works correctly with Llama Stack. + +## Troubleshooting + +If your external API isn't being loaded: + +1. Check that the `external_apis_dir` path is correct and accessible. +2. Verify that the YAML files are properly formatted. +3. Ensure all required Python packages are installed. +4. Check the Llama Stack server logs for any error messages - turn on debug logging to get more information using `LLAMA_STACK_LOGGING=all=debug`. +5. Verify that the API package is installed in your Python environment. diff --git a/versioned_docs/version-v0.2.23/concepts/apis/index.mdx b/versioned_docs/version-v0.2.23/concepts/apis/index.mdx new file mode 100644 index 0000000..6e699d1 --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/apis/index.mdx @@ -0,0 +1,28 @@ +--- +title: APIs +description: Available REST APIs and planned capabilities in Llama Stack +sidebar_label: APIs +sidebar_position: 1 +--- + +# APIs + +A Llama Stack API is described as a collection of REST endpoints. We currently support the following APIs: + +- **Inference**: run inference with a LLM +- **Safety**: apply safety policies to the output at a Systems (not only model) level +- **Agents**: run multi-step agentic workflows with LLMs with tool usage, memory (RAG), etc. +- **DatasetIO**: interface with datasets and data loaders +- **Scoring**: evaluate outputs of the system +- **Eval**: generate outputs (via Inference or Agents) and perform scoring +- **VectorIO**: perform operations on vector stores, such as adding documents, searching, and deleting documents +- **Telemetry**: collect telemetry data from the system +- **Post Training**: fine-tune a model +- **Tool Runtime**: interact with various tools and protocols +- **Responses**: generate responses from an LLM using this OpenAI compatible API. + +We are working on adding a few more APIs to complete the application lifecycle. These will include: +- **Batch Inference**: run inference on a dataset of inputs +- **Batch Agents**: run agents on a dataset of inputs +- **Synthetic Data Generation**: generate synthetic data for model development +- **Batches**: OpenAI-compatible batch management for inference diff --git a/versioned_docs/version-v0.2.23/concepts/architecture.mdx b/versioned_docs/version-v0.2.23/concepts/architecture.mdx new file mode 100644 index 0000000..8e97384 --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/architecture.mdx @@ -0,0 +1,74 @@ +--- +title: Llama Stack Architecture +description: Understanding Llama Stack's service-oriented design and benefits +sidebar_label: Architecture +sidebar_position: 2 +--- + +# Llama Stack architecture + +Llama Stack allows you to build different layers of distributions for your AI workloads using various SDKs and API providers. + +Llama Stack + +## Benefits of Llama stack + +### Current challenges in custom AI applications + +Building production AI applications today requires solving multiple challenges: + +**Infrastructure Complexity** + +- Running large language models efficiently requires specialized infrastructure. +- Different deployment scenarios (local development, cloud, edge) need different solutions. +- Moving from development to production often requires significant rework. + +**Essential Capabilities** + +- Safety guardrails and content filtering are necessary in an enterprise setting. +- Just model inference is not enough - Knowledge retrieval and RAG capabilities are required. +- Nearly any application needs composable multi-step workflows. +- Without monitoring, observability and evaluation, you end up operating in the dark. + +**Lack of Flexibility and Choice** + +- Directly integrating with multiple providers creates tight coupling. +- Different providers have different APIs and abstractions. +- Changing providers requires significant code changes. + +### Our Solution: A Universal Stack + +Llama Stack addresses these challenges through a service-oriented, API-first approach: + +**Develop Anywhere, Deploy Everywhere** +- Start locally with CPU-only setups +- Move to GPU acceleration when needed +- Deploy to cloud or edge without code changes +- Same APIs and developer experience everywhere + +**Production-Ready Building Blocks** +- Pre-built safety guardrails and content filtering +- Built-in RAG and agent capabilities +- Comprehensive evaluation toolkit +- Full observability and monitoring + +**True Provider Independence** +- Swap providers without application changes +- Mix and match best-in-class implementations +- Federation and fallback support +- No vendor lock-in + +**Robust Ecosystem** +- Llama Stack is already integrated with distribution partners (cloud providers, hardware vendors, and AI-focused companies). +- Ecosystem offers tailored infrastructure, software, and services for deploying a variety of models. + + +## Our Philosophy + +- **Service-Oriented**: REST APIs enforce clean interfaces and enable seamless transitions across different environments. +- **Composability**: Every component is independent but works together seamlessly +- **Production Ready**: Built for real-world applications, not just demos +- **Turnkey Solutions**: Easy to deploy built in solutions for popular deployment scenarios + + +With Llama Stack, you can focus on building your application while we handle the infrastructure complexity, essential capabilities, and provider integrations. diff --git a/versioned_docs/version-v0.2.23/concepts/distributions.mdx b/versioned_docs/version-v0.2.23/concepts/distributions.mdx new file mode 100644 index 0000000..5680996 --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/distributions.mdx @@ -0,0 +1,16 @@ +--- +title: Distributions +description: Pre-packaged provider configurations for different deployment scenarios +sidebar_label: Distributions +sidebar_position: 3 +--- + +# Distributions + +While there is a lot of flexibility to mix-and-match providers, often users will work with a specific set of providers (hardware support, contractual obligations, etc.) We therefore need to provide a _convenient shorthand_ for such collections. We call this shorthand a **Llama Stack Distribution** or a **Distro**. One can think of it as specific pre-packaged versions of the Llama Stack. Here are some examples: + +**Remotely Hosted Distro**: These are the simplest to consume from a user perspective. You can simply obtain the API key for these providers, point to a URL and have _all_ Llama Stack APIs working out of the box. Currently, [Fireworks](https://fireworks.ai/) and [Together](https://together.xyz/) provide such easy-to-consume Llama Stack distributions. + +**Locally Hosted Distro**: You may want to run Llama Stack on your own hardware. Typically though, you still need to use Inference via an external service. You can use providers like HuggingFace TGI, Fireworks, Together, etc. for this purpose. Or you may have access to GPUs and can run a [vLLM](https://github.com/vllm-project/vllm) or [NVIDIA NIM](https://build.nvidia.com/nim?filters=nimType%3Anim_type_run_anywhere&q=llama) instance. If you "just" have a regular desktop machine, you can use [Ollama](https://ollama.com/) for inference. To provide convenient quick access to these options, we provide a number of such pre-configured locally-hosted Distros. + +**On-device Distro**: To run Llama Stack directly on an edge device (mobile phone or a tablet), we provide Distros for [iOS](/docs/distributions/ondevice_distro/ios_sdk) and [Android](/docs/distributions/ondevice_distro/android_sdk) diff --git a/versioned_docs/version-v0.2.23/concepts/evaluation_concepts.mdx b/versioned_docs/version-v0.2.23/concepts/evaluation_concepts.mdx new file mode 100644 index 0000000..c7a13fd --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/evaluation_concepts.mdx @@ -0,0 +1,78 @@ +--- +title: Evaluation Concepts +description: Running evaluations on Llama Stack +sidebar_label: Evaluation Concepts +sidebar_position: 5 +--- + +# Evaluation Concepts + +The Llama Stack Evaluation flow allows you to run evaluations on your GenAI application datasets or pre-registered benchmarks. + +We introduce a set of APIs in Llama Stack for supporting running evaluations of LLM applications: +- `/datasetio` + `/datasets` API +- `/scoring` + `/scoring_functions` API +- `/eval` + `/benchmarks` API + +This guide goes over the sets of APIs and developer experience flow of using Llama Stack to run evaluations for different use cases. Checkout our Colab notebook on working examples with evaluations [here](https://colab.research.google.com/drive/10CHyykee9j2OigaIcRv47BKG9mrNm0tJ?usp=sharing). + +The Evaluation APIs are associated with a set of Resources. Please visit the Resources section in our [Core Concepts](./index.mdx) guide for better high-level understanding. + +- **DatasetIO**: defines interface with datasets and data loaders. + - Associated with `Dataset` resource. +- **Scoring**: evaluate outputs of the system. + - Associated with `ScoringFunction` resource. We provide a suite of out-of-the box scoring functions and also the ability for you to add custom evaluators. These scoring functions are the core part of defining an evaluation task to output evaluation metrics. +- **Eval**: generate outputs (via Inference or Agents) and perform scoring. + - Associated with `Benchmark` resource. + +## Open-benchmark Eval + +### List of open-benchmarks Llama Stack support + +Llama stack pre-registers several popular open-benchmarks to easily evaluate model perfomance via CLI. + +The list of open-benchmarks we currently support: +- [MMLU-COT](https://arxiv.org/abs/2009.03300) (Measuring Massive Multitask Language Understanding): Benchmark designed to comprehensively evaluate the breadth and depth of a model's academic and professional understanding +- [GPQA-COT](https://arxiv.org/abs/2311.12022) (A Graduate-Level Google-Proof Q&A Benchmark): A challenging benchmark of 448 multiple-choice questions written by domain experts in biology, physics, and chemistry. +- [SimpleQA](https://openai.com/index/introducing-simpleqa/): Benchmark designed to access models to answer short, fact-seeking questions. +- [MMMU](https://arxiv.org/abs/2311.16502) (A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI)]: Benchmark designed to evaluate multimodal models. + +You can follow this [contributing guide](../references/evals_reference/#open-benchmark-contributing-guide) to add more open-benchmarks to Llama Stack + +### Run evaluation on open-benchmarks via CLI + +We have built-in functionality to run the supported open-benckmarks using llama-stack-client CLI + +#### Spin up Llama Stack server + +Spin up llama stack server with 'open-benchmark' template +```bash +llama stack run llama_stack/distributions/open-benchmark/run.yaml +``` + +#### Run eval CLI +There are 3 necessary inputs to run a benchmark eval +- `list of benchmark_ids`: The list of benchmark ids to run evaluation on +- `model-id`: The model id to evaluate on +- `output_dir`: Path to store the evaluate results + +```bash +llama-stack-client eval run-benchmark ... \ +--model_id \ +--output_dir +``` + +You can run +```bash +llama-stack-client eval run-benchmark help +``` +to see the description of all the flags that eval run-benchmark has + +In the output log, you can find the file path that has your evaluation results. Open that file and you can see you aggregate +evaluation results over there. + +## What's Next? + +- Check out our Colab notebook on working examples with running benchmark evaluations [here](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb#scrollTo=mxLCsP4MvFqP). +- Check out our [Building Applications - Evaluation](../building_applications/evals.mdx) guide for more details on how to use the Evaluation APIs to evaluate your applications. +- Check out our [Evaluation Reference](../references/evals_reference/) for more details on the APIs. diff --git a/versioned_docs/version-v0.2.23/concepts/index.mdx b/versioned_docs/version-v0.2.23/concepts/index.mdx new file mode 100644 index 0000000..1278ef9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/index.mdx @@ -0,0 +1,31 @@ +--- +title: Core Concepts +description: Understanding Llama Stack's service-oriented philosophy and key concepts +sidebar_label: Overview +sidebar_position: 1 +--- + +Given Llama Stack's service-oriented philosophy, a few concepts and workflows arise which may not feel completely natural in the LLM landscape, especially if you are coming with a background in other frameworks. + +## Documentation Structure + +This section covers the fundamental concepts of Llama Stack: + +- **[Architecture](architecture.mdx)** - Learn about Llama Stack's architectural design and principles +- **[APIs](/docs/concepts/apis/)** - Understanding the core APIs and their stability levels + - [API Overview](apis/index.mdx) - Core APIs available in Llama Stack + - [API Providers](apis/api_providers.mdx) - How providers implement APIs + - [External APIs](apis/external.mdx) - External APIs available in Llama Stack + - [API Stability Leveling](apis/api_leveling.mdx) - API stability and versioning +- **[Distributions](distributions.mdx)** - Pre-configured deployment packages +- **[Resources](resources.mdx)** - Understanding Llama Stack resources and their lifecycle + +## Getting Started + +If you're new to Llama Stack, we recommend starting with: + +1. **[Architecture](architecture.mdx)** - Understand the overall system design +2. **[APIs](apis/index.mdx)** - Learn about the available APIs and their purpose +3. **[Distributions](distributions.mdx)** - Choose a pre-configured setup for your use case + +Each concept builds upon the previous ones to give you a comprehensive understanding of how Llama Stack works and how to use it effectively. diff --git a/versioned_docs/version-v0.2.23/concepts/resources.mdx b/versioned_docs/version-v0.2.23/concepts/resources.mdx new file mode 100644 index 0000000..8d1bd22 --- /dev/null +++ b/versioned_docs/version-v0.2.23/concepts/resources.mdx @@ -0,0 +1,26 @@ +--- +title: Resources +description: Resource federation and registration in Llama Stack +sidebar_label: Resources +sidebar_position: 4 +--- + +# Resources + +Some of these APIs are associated with a set of **Resources**. Here is the mapping of APIs to resources: + +- **Inference**, **Eval** and **Post Training** are associated with `Model` resources. +- **Safety** is associated with `Shield` resources. +- **Tool Runtime** is associated with `ToolGroup` resources. +- **DatasetIO** is associated with `Dataset` resources. +- **VectorIO** is associated with `VectorDB` resources. +- **Scoring** is associated with `ScoringFunction` resources. +- **Eval** is associated with `Model` and `Benchmark` resources. + +Furthermore, we allow these resources to be **federated** across multiple providers. For example, you may have some Llama models served by Fireworks while others are served by AWS Bedrock. Regardless, they will all work seamlessly with the same uniform Inference API provided by Llama Stack. + +:::tip Registering Resources + +Given this architecture, it is necessary for the Stack to know which provider to use for a given resource. This means you need to explicitly _register_ resources (including models) before you can use them with the associated APIs. + +::: diff --git a/versioned_docs/version-v0.2.23/contributing/index.mdx b/versioned_docs/version-v0.2.23/contributing/index.mdx new file mode 100644 index 0000000..263900e --- /dev/null +++ b/versioned_docs/version-v0.2.23/contributing/index.mdx @@ -0,0 +1,233 @@ +# Contributing to Llama Stack +We want to make contributing to this project as easy and transparent as +possible. + +## Set up your development environment + +We use [uv](https://github.com/astral-sh/uv) to manage python dependencies and virtual environments. +You can install `uv` by following this [guide](https://docs.astral.sh/uv/getting-started/installation/). + +You can install the dependencies by running: + +```bash +cd llama-stack +uv sync --group dev +uv pip install -e . +source .venv/bin/activate +``` + +```{note} +You can use a specific version of Python with `uv` by adding the `--python ` flag (e.g. `--python 3.12`). +Otherwise, `uv` will automatically select a Python version according to the `requires-python` section of the `pyproject.toml`. +For more info, see the [uv docs around Python versions](https://docs.astral.sh/uv/concepts/python-versions/). +``` + +Note that you can create a dotenv file `.env` that includes necessary environment variables: +``` +LLAMA_STACK_BASE_URL=http://localhost:8321 +LLAMA_STACK_CLIENT_LOG=debug +LLAMA_STACK_PORT=8321 +LLAMA_STACK_CONFIG= +TAVILY_SEARCH_API_KEY= +BRAVE_SEARCH_API_KEY= +``` + +And then use this dotenv file when running client SDK tests via the following: +```bash +uv run --env-file .env -- pytest -v tests/integration/inference/test_text_inference.py --text-model=meta-llama/Llama-3.1-8B-Instruct +``` + +### Pre-commit Hooks + +We use [pre-commit](https://pre-commit.com/) to run linting and formatting checks on your code. You can install the pre-commit hooks by running: + +```bash +uv run pre-commit install +``` + +After that, pre-commit hooks will run automatically before each commit. + +Alternatively, if you don't want to install the pre-commit hooks, you can run the checks manually by running: + +```bash +uv run pre-commit run --all-files +``` + +```{caution} +Before pushing your changes, make sure that the pre-commit hooks have passed successfully. +``` + +## Discussions -> Issues -> Pull Requests + +We actively welcome your pull requests. However, please read the following. This is heavily inspired by [Ghostty](https://github.com/ghostty-org/ghostty/blob/main/CONTRIBUTING.md). + +If in doubt, please open a [discussion](https://github.com/meta-llama/llama-stack/discussions); we can always convert that to an issue later. + +### Issues +We use GitHub issues to track public bugs. Please ensure your description is +clear and has sufficient instructions to be able to reproduce the issue. + +Meta has a [bounty program](http://facebook.com/whitehat/info) for the safe +disclosure of security bugs. In those cases, please go through the process +outlined on that page and do not file a public issue. + +### Contributor License Agreement ("CLA") +In order to accept your pull request, we need you to submit a CLA. You only need +to do this once to work on any of Meta's open source projects. + +Complete your CLA here: [https://code.facebook.com/cla](https://code.facebook.com/cla) + +**I'd like to contribute!** + +If you are new to the project, start by looking at the issues tagged with "good first issue". If you're interested +leave a comment on the issue and a triager will assign it to you. + +Please avoid picking up too many issues at once. This helps you stay focused and ensures that others in the community also have opportunities to contribute. +- Try to work on only 1โ€“2 issues at a time, especially if youโ€™re still getting familiar with the codebase. +- Before taking an issue, check if itโ€™s already assigned or being actively discussed. +- If youโ€™re blocked or canโ€™t continue with an issue, feel free to unassign yourself or leave a comment so others can step in. + +**I have a bug!** + +1. Search the issue tracker and discussions for similar issues. +2. If you don't have steps to reproduce, open a discussion. +3. If you have steps to reproduce, open an issue. + +**I have an idea for a feature!** + +1. Open a discussion. + +**I've implemented a feature!** + +1. If there is an issue for the feature, open a pull request. +2. If there is no issue, open a discussion and link to your branch. + +**I have a question!** + +1. Open a discussion or use [Discord](https://discord.gg/llama-stack). + + +**Opening a Pull Request** + +1. Fork the repo and create your branch from `main`. +2. If you've changed APIs, update the documentation. +3. Ensure the test suite passes. +4. Make sure your code lints using `pre-commit`. +5. If you haven't already, complete the Contributor License Agreement ("CLA"). +6. Ensure your pull request follows the [conventional commits format](https://www.conventionalcommits.org/en/v1.0.0/). +7. Ensure your pull request follows the [coding style](#coding-style). + + +Please keep pull requests (PRs) small and focused. If you have a large set of changes, consider splitting them into logically grouped, smaller PRs to facilitate review and testing. + +```{tip} +As a general guideline: +- Experienced contributors should try to keep no more than 5 open PRs at a time. +- New contributors are encouraged to have only one open PR at a time until theyโ€™re familiar with the codebase and process. +``` + +## Repository guidelines + +### Coding Style + +* Comments should provide meaningful insights into the code. Avoid filler comments that simply + describe the next step, as they create unnecessary clutter, same goes for docstrings. +* Prefer comments to clarify surprising behavior and/or relationships between parts of the code + rather than explain what the next line of code does. +* Catching exceptions, prefer using a specific exception type rather than a broad catch-all like + `Exception`. +* Error messages should be prefixed with "Failed to ..." +* 4 spaces for indentation rather than tab +* When using `# noqa` to suppress a style or linter warning, include a comment explaining the + justification for bypassing the check. +* When using `# type: ignore` to suppress a mypy warning, include a comment explaining the + justification for bypassing the check. +* Don't use unicode characters in the codebase. ASCII-only is preferred for compatibility or + readability reasons. +* Providers configuration class should be Pydantic Field class. It should have a `description` field + that describes the configuration. These descriptions will be used to generate the provider + documentation. +* When possible, use keyword arguments only when calling functions. +* Llama Stack utilizes custom Exception classes for certain Resources that should be used where applicable. + +### License +By contributing to Llama, you agree that your contributions will be licensed +under the LICENSE file in the root directory of this source tree. + +## Common Tasks + +Some tips about common tasks you work on while contributing to Llama Stack: + +### Using `llama stack build` + +Building a stack image will use the production version of the `llama-stack` and `llama-stack-client` packages. If you are developing with a llama-stack repository checked out and need your code to be reflected in the stack image, set `LLAMA_STACK_DIR` and `LLAMA_STACK_CLIENT_DIR` to the appropriate checked out directories when running any of the `llama` CLI commands. + +Example: +```bash +cd work/ +git clone https://github.com/meta-llama/llama-stack.git +git clone https://github.com/meta-llama/llama-stack-client-python.git +cd llama-stack +LLAMA_STACK_DIR=$(pwd) LLAMA_STACK_CLIENT_DIR=../llama-stack-client-python llama stack build --distro <...> +``` + +### Updating distribution configurations + +If you have made changes to a provider's configuration in any form (introducing a new config key, or +changing models, etc.), you should run `./scripts/distro_codegen.py` to re-generate various YAML +files as well as the documentation. You should not change `docs/source/.../distributions/` files +manually as they are auto-generated. + +### Updating the provider documentation + +If you have made changes to a provider's configuration, you should run `./scripts/provider_codegen.py` +to re-generate the documentation. You should not change `docs/source/.../providers/` files manually +as they are auto-generated. +Note that the provider "description" field will be used to generate the provider documentation. + +### Building the Documentation + +If you are making changes to the documentation at [https://llamastack.github.io/](https://llamastack.github.io/), you can use the following command to build the documentation and preview your changes. + +```bash +# This rebuilds the documentation pages and the OpenAPI spec. +npm install +npm run gen-api-docs all +npm run build + +# This will start a local server (usually at http://127.0.0.1:3000). +npm run serve +``` + +### Update API Documentation + +If you modify or add new API endpoints, update the API documentation accordingly. You can do this by running the following command: + +```bash +uv run ./docs/openapi_generator/run_openapi_generator.sh +``` + +The generated API schema will be available in `docs/static/`. Make sure to review the changes before committing. + +## Adding a New Provider + +See: +- [Adding a New API Provider Page](./new_api_provider.mdx) which describes how to add new API providers to the Stack. +- [Vector Database Page](./new_vector_database.mdx) which describes how to add a new vector databases with Llama Stack. +- [External Provider Page](/docs/providers/external/) which describes how to add external providers to the Stack. + + +## Testing + + +See the [Testing README](https://github.com/meta-llama/llama-stack/blob/main/tests/README.md) for detailed testing information. + +## Advanced Topics + +For developers who need deeper understanding of the testing system internals: + +- [Record-Replay Testing](./testing/record-replay.mdx) + +### Benchmarking + +See the [Benchmarking README](https://github.com/meta-llama/llama-stack/blob/main/benchmarking/k8s-benchmark/README.md) for benchmarking information. diff --git a/versioned_docs/version-v0.2.23/contributing/new_api_provider.mdx b/versioned_docs/version-v0.2.23/contributing/new_api_provider.mdx new file mode 100644 index 0000000..4ae6d5e --- /dev/null +++ b/versioned_docs/version-v0.2.23/contributing/new_api_provider.mdx @@ -0,0 +1,98 @@ +--- +title: Adding a New API Provider +description: Guide for adding new API providers to Llama Stack +sidebar_label: New API Provider +sidebar_position: 2 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +This guide will walk you through the process of adding a new API provider to Llama Stack. + + +- Begin by reviewing the [core concepts](../concepts/) of Llama Stack and choose the API your provider belongs to (Inference, Safety, VectorIO, etc.) +- Determine the provider type ([Remote](https://github.com/meta-llama/llama-stack/tree/main/llama_stack/providers/remote) or [Inline](https://github.com/meta-llama/llama-stack/tree/main/llama_stack/providers/inline)). Remote providers make requests to external services, while inline providers execute implementation locally. +- Add your provider to the appropriate [Registry](https://github.com/meta-llama/llama-stack/tree/main/llama_stack/providers/registry/). Specify pip dependencies necessary. +- Update any distribution [Templates](https://github.com/meta-llama/llama-stack/tree/main/llama_stack/distributions/) `build.yaml` and `run.yaml` files if they should include your provider by default. Run [./scripts/distro_codegen.py](https://github.com/meta-llama/llama-stack/blob/main/scripts/distro_codegen.py) if necessary. Note that `distro_codegen.py` will fail if the new provider causes any distribution template to attempt to import provider-specific dependencies. This usually means the distribution's `get_distribution_template()` code path should only import any necessary Config or model alias definitions from each provider and not the provider's actual implementation. + + +Here are some example PRs to help you get started: + - [Grok Inference Implementation](https://github.com/meta-llama/llama-stack/pull/609) + - [Nvidia Inference Implementation](https://github.com/meta-llama/llama-stack/pull/355) + - [Model context protocol Tool Runtime](https://github.com/meta-llama/llama-stack/pull/665) + +## Guidelines for creating Internal or External Providers + +|**Type** |Internal (In-tree) |External (out-of-tree) +|---------|-------------------|---------------------| +|**Description** |A provider that is directly in the Llama Stack code|A provider that is outside of the Llama stack core codebase but is still accessible and usable by Llama Stack. +|**Benefits** |Ability to interact with the provider with minimal additional configurations or installations| Contributors do not have to add directly to the code to create providers accessible on Llama Stack. Keep provider-specific code separate from the core Llama Stack code. + +## Inference Provider Patterns + +When implementing Inference providers for OpenAI-compatible APIs, Llama Stack provides several mixin classes to simplify development and ensure consistent behavior across providers. + +### OpenAIMixin + +The `OpenAIMixin` class provides direct OpenAI API functionality for providers that work with OpenAI-compatible endpoints. It includes: + +#### Direct API Methods +- **`openai_completion()`**: Legacy text completion API with full parameter support +- **`openai_chat_completion()`**: Chat completion API supporting streaming, tools, and function calling +- **`openai_embeddings()`**: Text embeddings generation with customizable encoding and dimensions + +#### Model Management +- **`check_model_availability()`**: Queries the API endpoint to verify if a model exists and is accessible + +#### Client Management +- **`client` property**: Automatically creates and configures AsyncOpenAI client instances using your provider's credentials + +#### Required Implementation + +To use `OpenAIMixin`, your provider must implement these abstract methods: + +```python +@abstractmethod +def get_api_key(self) -> str: + """Return the API key for authentication""" + pass + + +@abstractmethod +def get_base_url(self) -> str: + """Return the OpenAI-compatible API base URL""" + pass +``` + +## Testing the Provider + +Before running tests, you must have required dependencies installed. This depends on the providers or distributions you are testing. For example, if you are testing the `together` distribution, you should install dependencies via `llama stack build --distro together`. + +### 1. Integration Testing + +Integration tests are located in [tests/integration](https://github.com/meta-llama/llama-stack/tree/main/tests/integration). These tests use the python client-SDK APIs (from the `llama_stack_client` package) to test functionality. Since these tests use client APIs, they can be run either by pointing to an instance of the Llama Stack server or "inline" by using `LlamaStackAsLibraryClient`. + +Consult [tests/integration/README.md](https://github.com/meta-llama/llama-stack/blob/main/tests/integration/README.md) for more details on how to run the tests. + +Note that each provider's `sample_run_config()` method (in the configuration class for that provider) + typically references some environment variables for specifying API keys and the like. You can set these in the environment or pass these via the `--env` flag to the test command. + + +### 2. Unit Testing + +Unit tests are located in [tests/unit](https://github.com/meta-llama/llama-stack/tree/main/tests/unit). Provider-specific unit tests are located in [tests/unit/providers](https://github.com/meta-llama/llama-stack/tree/main/tests/unit/providers). These tests are all run automatically as part of the CI process. + +Consult [tests/unit/README.md](https://github.com/meta-llama/llama-stack/blob/main/tests/unit/README.md) for more details on how to run the tests manually. + +### 3. Additional end-to-end testing + +1. Start a Llama Stack server with your new provider +2. Verify compatibility with existing client scripts in the [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main) repository +3. Document which scripts are compatible with your provider + +## Submitting Your PR + +1. Ensure all tests pass +2. Include a comprehensive test plan in your PR summary +3. Document any known limitations or considerations diff --git a/versioned_docs/version-v0.2.23/contributing/new_vector_database.mdx b/versioned_docs/version-v0.2.23/contributing/new_vector_database.mdx new file mode 100644 index 0000000..044e2f6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/contributing/new_vector_database.mdx @@ -0,0 +1,83 @@ +--- +title: Adding a New Vector Database +description: Guide for adding new vector database providers to Llama Stack +sidebar_label: New Vector Database +sidebar_position: 3 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +This guide will walk you through the process of adding a new vector database to Llama Stack. + +> **_NOTE:_** Here's an example Pull Request of the [Milvus Vector Database Provider](https://github.com/meta-llama/llama-stack/pull/1467). + +Vector Database providers are used to store and retrieve vector embeddings. Vector databases are not limited to vector +search but can support keyword and hybrid search. Additionally, vector database can also support operations like +filtering, sorting, and aggregating vectors. + +## Steps to Add a New Vector Database Provider +1. **Choose the Database Type**: Determine if your vector database is a remote service, inline, or both. + - Remote databases make requests to external services, while inline databases execute locally. Some providers support both. +2. **Implement the Provider**: Create a new provider class that inherits from `VectorDatabaseProvider` and implements the required methods. + - Implement methods for vector storage, retrieval, search, and any additional features your database supports. + - You will need to implement the following methods for `YourVectorIndex`: + - `YourVectorIndex.create()` + - `YourVectorIndex.initialize()` + - `YourVectorIndex.add_chunks()` + - `YourVectorIndex.delete_chunk()` + - `YourVectorIndex.query_vector()` + - `YourVectorIndex.query_keyword()` + - `YourVectorIndex.query_hybrid()` + - You will need to implement the following methods for `YourVectorIOAdapter`: + - `YourVectorIOAdapter.initialize()` + - `YourVectorIOAdapter.shutdown()` + - `YourVectorIOAdapter.list_vector_dbs()` + - `YourVectorIOAdapter.register_vector_db()` + - `YourVectorIOAdapter.unregister_vector_db()` + - `YourVectorIOAdapter.insert_chunks()` + - `YourVectorIOAdapter.query_chunks()` + - `YourVectorIOAdapter.delete_chunks()` +3. **Add to Registry**: Register your provider in the appropriate registry file. + - Update [llama_stack/providers/registry/vector_io.py](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/registry/vector_io.py) to include your new provider. +```python +from llama_stack.providers.registry.specs import InlineProviderSpec +from llama_stack.providers.registry.api import Api + +InlineProviderSpec( + api=Api.vector_io, + provider_type="inline::milvus", + pip_packages=["pymilvus>=2.4.10"], + module="llama_stack.providers.inline.vector_io.milvus", + config_class="llama_stack.providers.inline.vector_io.milvus.MilvusVectorIOConfig", + api_dependencies=[Api.inference], + optional_api_dependencies=[Api.files], + description="", +), +``` +4. **Add Tests**: Create unit tests and integration tests for your provider in the `tests/` directory. + - Unit Tests + - By following the structure of the class methods, you will be able to easily run unit and integration tests for your database. + 1. You have to configure the tests for your provide in `/tests/unit/providers/vector_io/conftest.py`. + 2. Update the `vector_provider` fixture to include your provider if they are an inline provider. + 3. Create a `your_vectorprovider_index` fixture that initializes your vector index. + 4. Create a `your_vectorprovider_adapter` fixture that initializes your vector adapter. + 5. Add your provider to the `vector_io_providers` fixture dictionary. + - Please follow the naming convention of `your_vectorprovider_index` and `your_vectorprovider_adapter` as the tests require this to execute properly. + - Integration Tests + - Integration tests are located in [tests/integration](https://github.com/meta-llama/llama-stack/tree/main/tests/integration). These tests use the python client-SDK APIs (from the `llama_stack_client` package) to test functionality. + - The two set of integration tests are: + - `tests/integration/vector_io/test_vector_io.py`: This file tests registration, insertion, and retrieval. + - `tests/integration/vector_io/test_openai_vector_stores.py`: These tests are for OpenAI-compatible vector stores and test the OpenAI API compatibility. + - You will need to update `skip_if_provider_doesnt_support_openai_vector_stores` to include your provider as well as `skip_if_provider_doesnt_support_openai_vector_stores_search` to test the appropriate search functionality. + - Running the tests in the GitHub CI + - You will need to update the `.github/workflows/integration-vector-io-tests.yml` file to include your provider. + - If your provider is a remote provider, you will also have to add a container to spin up and run it in the action. + - Updating the pyproject.yml + - If you are adding tests for the `inline` provider you will have to update the `unit` group. + - `uv add new_pip_package --group unit` + - If you are adding tests for the `remote` provider you will have to update the `test` group, which is used in the GitHub CI for integration tests. + - `uv add new_pip_package --group test` +5. **Update Documentation**: Please update the documentation for end users + - Generate the provider documentation by running [./scripts/provider_codegen.py](https://github.com/meta-llama/llama-stack/blob/main/scripts/provider_codegen.py). + - Update the autogenerated content in the registry/vector_io.py file with information about your provider. Please see other providers for examples. diff --git a/versioned_docs/version-v0.2.23/contributing/testing/record-replay.mdx b/versioned_docs/version-v0.2.23/contributing/testing/record-replay.mdx new file mode 100644 index 0000000..47803c1 --- /dev/null +++ b/versioned_docs/version-v0.2.23/contributing/testing/record-replay.mdx @@ -0,0 +1,241 @@ +--- +title: Record-Replay Testing System +description: Understanding how Llama Stack captures and replays API interactions for testing +sidebar_label: Record-Replay System +sidebar_position: 4 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Record-Replay System + +Understanding how Llama Stack captures and replays API interactions for testing. + +## Overview + +The record-replay system solves a fundamental challenge in AI testing: how do you test against expensive, non-deterministic APIs without breaking the bank or dealing with flaky tests? + +The solution: intercept API calls, store real responses, and replay them later. This gives you real API behavior without the cost or variability. + +## How It Works + +### Request Hashing + +Every API request gets converted to a deterministic hash for lookup: + +```python +def normalize_request(method: str, url: str, headers: dict, body: dict) -> str: + normalized = { + "method": method.upper(), + "endpoint": urlparse(url).path, # Just the path, not full URL + "body": body, # Request parameters + } + return hashlib.sha256(json.dumps(normalized, sort_keys=True).encode()).hexdigest() +``` + +**Key insight:** The hashing is intentionally precise. Different whitespace, float precision, or parameter order produces different hashes. This prevents subtle bugs from false cache hits. + +```python +# These produce DIFFERENT hashes: +{"content": "Hello world"} +{"content": "Hello world\n"} +{"temperature": 0.7} +{"temperature": 0.7000001} +``` + +### Client Interception + +The system patches OpenAI and Ollama client methods to intercept calls before they leave your application. This happens transparently - your test code doesn't change. + +### Storage Architecture + +Recordings are stored as JSON files in the recording directory. They are looked up by their request hash. + +``` +recordings/ +โ””โ”€โ”€ responses/ + โ”œโ”€โ”€ abc123def456.json # Individual response files + โ””โ”€โ”€ def789ghi012.json +``` + +**JSON files** store complete request/response pairs in human-readable format for debugging. + +## Recording Modes + +### LIVE Mode + +Direct API calls with no recording or replay: + +```python +with inference_recording(mode=InferenceMode.LIVE): + response = await client.chat.completions.create(...) +``` + +Use for initial development and debugging against real APIs. + +### RECORD Mode + +Captures API interactions while passing through real responses: + +```python +with inference_recording(mode=InferenceMode.RECORD, storage_dir="./recordings"): + response = await client.chat.completions.create(...) + # Real API call made, response captured AND returned +``` + +The recording process: +1. Request intercepted and hashed +2. Real API call executed +3. Response captured and serialized +4. Recording stored to disk +5. Original response returned to caller + +### REPLAY Mode + +Returns stored responses instead of making API calls: + +```python +with inference_recording(mode=InferenceMode.REPLAY, storage_dir="./recordings"): + response = await client.chat.completions.create(...) + # No API call made, cached response returned instantly +``` + +The replay process: +1. Request intercepted and hashed +2. Hash looked up in SQLite index +3. Response loaded from JSON file +4. Response deserialized and returned +5. Error if no recording found + +## Streaming Support + +Streaming APIs present a unique challenge: how do you capture an async generator? + +### The Problem + +```python +# How do you record this? +async for chunk in client.chat.completions.create(stream=True): + process(chunk) +``` + +### The Solution + +The system captures all chunks immediately before yielding any: + +```python +async def handle_streaming_record(response): + # Capture complete stream first + chunks = [] + async for chunk in response: + chunks.append(chunk) + + # Store complete recording + storage.store_recording( + request_hash, request_data, {"body": chunks, "is_streaming": True} + ) + + # Return generator that replays captured chunks + async def replay_stream(): + for chunk in chunks: + yield chunk + + return replay_stream() +``` + +This ensures: +- **Complete capture** - The entire stream is saved atomically +- **Interface preservation** - The returned object behaves like the original API +- **Deterministic replay** - Same chunks in the same order every time + +## Serialization + +API responses contain complex Pydantic objects that need careful serialization: + +```python +def _serialize_response(response): + if hasattr(response, "model_dump"): + # Preserve type information for proper deserialization + return { + "__type__": f"{response.__class__.__module__}.{response.__class__.__qualname__}", + "__data__": response.model_dump(mode="json"), + } + return response +``` + +This preserves type safety - when replayed, you get the same Pydantic objects with all their validation and methods. + +## Environment Integration + +### Environment Variables + +Control recording behavior globally: + +```bash +export LLAMA_STACK_TEST_INFERENCE_MODE=replay # this is the default +export LLAMA_STACK_TEST_RECORDING_DIR=/path/to/recordings # default is tests/integration/recordings +pytest tests/integration/ +``` + +### Pytest Integration + +The system integrates automatically based on environment variables, requiring no changes to test code. + +## Debugging Recordings + +### Inspecting Storage + +```bash +# See what's recorded +sqlite3 recordings/index.sqlite "SELECT endpoint, model, timestamp FROM recordings LIMIT 10;" + +# View specific response +cat recordings/responses/abc123def456.json | jq '.response.body' + +# Find recordings by endpoint +sqlite3 recordings/index.sqlite "SELECT * FROM recordings WHERE endpoint='/v1/chat/completions';" +``` + +### Common Issues + +**Hash mismatches:** Request parameters changed slightly between record and replay +```bash +# Compare request details +cat recordings/responses/abc123.json | jq '.request' +``` + +**Serialization errors:** Response types changed between versions +```bash +# Re-record with updated types +rm recordings/responses/failing_hash.json +LLAMA_STACK_TEST_INFERENCE_MODE=record pytest test_failing.py +``` + +**Missing recordings:** New test or changed parameters +```bash +# Record the missing interaction +LLAMA_STACK_TEST_INFERENCE_MODE=record pytest test_new.py +``` + +## Design Decisions + +### Why Not Mocks? + +Traditional mocking breaks down with AI APIs because: +- Response structures are complex and evolve frequently +- Streaming behavior is hard to mock correctly +- Edge cases in real APIs get missed +- Mocks become brittle maintenance burdens + +### Why Precise Hashing? + +Loose hashing (normalizing whitespace, rounding floats) seems convenient but hides bugs. If a test changes slightly, you want to know about it rather than accidentally getting the wrong cached response. + +### Why JSON + SQLite? + +- **JSON** - Human readable, diff-friendly, easy to inspect and modify +- **SQLite** - Fast indexed lookups without loading response bodies +- **Hybrid** - Best of both worlds for different use cases + +This system provides reliable, fast testing against real AI APIs while maintaining the ability to debug issues when they arise. diff --git a/versioned_docs/version-v0.2.23/deploying/aws_eks_deployment.mdx b/versioned_docs/version-v0.2.23/deploying/aws_eks_deployment.mdx new file mode 100644 index 0000000..fa107ea --- /dev/null +++ b/versioned_docs/version-v0.2.23/deploying/aws_eks_deployment.mdx @@ -0,0 +1,30 @@ +--- +title: AWS EKS Deployment Guide +description: Deploy Llama Stack on AWS EKS +sidebar_label: AWS EKS Deployment +sidebar_position: 3 +--- + +## AWS EKS Deployment + +### Prerequisites + +- Set up an [EKS cluster](https://docs.aws.amazon.com/eks/latest/userguide/getting-started.html) +- Create a [GitHub OAuth app](https://docs.github.com/en/apps/oauth-apps/building-oauth-apps/creating-an-oauth-app) +- Set authorization callback URL to `http:///api/auth/callback/` + +### Automated Deployment + +```bash +export HF_TOKEN= +export GITHUB_CLIENT_ID= +export GITHUB_CLIENT_SECRET= +export LLAMA_STACK_UI_URL= + +cd docs/source/distributions/eks +./apply.sh +``` + +This script will: +- Set up default storage class for AWS EKS +- Deploy Llama Stack server in Kubernetes pods and services diff --git a/versioned_docs/version-v0.2.23/deploying/index.mdx b/versioned_docs/version-v0.2.23/deploying/index.mdx new file mode 100644 index 0000000..eaa0e26 --- /dev/null +++ b/versioned_docs/version-v0.2.23/deploying/index.mdx @@ -0,0 +1,14 @@ +--- +title: Deploying Llama Stack +description: Production deployment guides for Llama Stack in various environments +sidebar_label: Overview +sidebar_position: 1 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Deploying Llama Stack + +[**โ†’ Kubernetes Deployment Guide**](./kubernetes_deployment.mdx) +[**โ†’ AWS EKS Deployment Guide**](./aws_eks_deployment.mdx) diff --git a/versioned_docs/version-v0.2.23/deploying/kubernetes_deployment.mdx b/versioned_docs/version-v0.2.23/deploying/kubernetes_deployment.mdx new file mode 100644 index 0000000..a937ce3 --- /dev/null +++ b/versioned_docs/version-v0.2.23/deploying/kubernetes_deployment.mdx @@ -0,0 +1,224 @@ +--- +title: Kubernetes Deployment Guide +description: Deploy Llama Stack on Kubernetes clusters with vLLM inference service +sidebar_label: Kubernetes +sidebar_position: 2 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +# Kubernetes Deployment Guide + +Deploy Llama Stack and vLLM servers in a Kubernetes cluster instead of running them locally. This guide covers both local development with Kind and production deployment on AWS EKS. + +## Prerequisites + +### Local Kubernetes Setup + +Create a local Kubernetes cluster via Kind: + +```bash +kind create cluster --image kindest/node:v1.32.0 --name llama-stack-test +``` + +Set your Hugging Face token: + +```bash +export HF_TOKEN=$(echo -n "your-hf-token" | base64) +``` + +## Quick Deployment + +### Step 1: Create Storage and Secrets + +```yaml +cat <$tmp_dir/Containerfile.llama-stack-run-k8s <-build.yaml` and template file `-run.yaml` will be generated and saved at the output file path specified at the end of the command. + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + + + +To build from alternative API providers, we provide distribution templates for users to get started building a distribution backed by different providers. + +The following command will allow you to see the available templates and their corresponding providers. +``` +llama stack build --list-templates +``` + +``` +------------------------------+-----------------------------------------------------------------------------+ +| Template Name | Description | ++------------------------------+-----------------------------------------------------------------------------+ +| watsonx | Use watsonx for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| vllm-gpu | Use a built-in vLLM engine for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| together | Use Together.AI for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| tgi | Use (an external) TGI server for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| starter | Quick start template for running Llama Stack with several popular providers | ++------------------------------+-----------------------------------------------------------------------------+ +| sambanova | Use SambaNova for running LLM inference and safety | ++------------------------------+-----------------------------------------------------------------------------+ +| remote-vllm | Use (an external) vLLM server for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| postgres-demo | Quick start template for running Llama Stack with several popular providers | ++------------------------------+-----------------------------------------------------------------------------+ +| passthrough | Use Passthrough hosted llama-stack endpoint for LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| open-benchmark | Distribution for running open benchmarks | ++------------------------------+-----------------------------------------------------------------------------+ +| ollama | Use (an external) Ollama server for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| nvidia | Use NVIDIA NIM for running LLM inference, evaluation and safety | ++------------------------------+-----------------------------------------------------------------------------+ +| meta-reference-gpu | Use Meta Reference for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| llama_api | Distribution for running e2e tests in CI | ++------------------------------+-----------------------------------------------------------------------------+ +| hf-serverless | Use (an external) Hugging Face Inference Endpoint for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| hf-endpoint | Use (an external) Hugging Face Inference Endpoint for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| groq | Use Groq for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| fireworks | Use Fireworks.AI for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| experimental-post-training | Experimental template for post training | ++------------------------------+-----------------------------------------------------------------------------+ +| dell | Dell's distribution of Llama Stack. TGI inference via Dell's custom | +| | container | ++------------------------------+-----------------------------------------------------------------------------+ +| ci-tests | Distribution for running e2e tests in CI | ++------------------------------+-----------------------------------------------------------------------------+ +| cerebras | Use Cerebras for running LLM inference | ++------------------------------+-----------------------------------------------------------------------------+ +| bedrock | Use AWS Bedrock for running LLM inference and safety | ++------------------------------+-----------------------------------------------------------------------------+ +``` + +You may then pick a template to build your distribution with providers fitted to your liking. + +For example, to build a distribution with TGI as the inference provider, you can run: +``` +$ llama stack build --distro starter +... +You can now edit ~/.llama/distributions/llamastack-starter/starter-run.yaml and run `llama stack run ~/.llama/distributions/llamastack-starter/starter-run.yaml` +``` + +```{tip} +The generated `run.yaml` file is a starting point for your configuration. For comprehensive guidance on customizing it for your specific needs, infrastructure, and deployment scenarios, see [Customizing Your run.yaml Configuration](customizing_run_yaml.md). +``` + + + +If the provided templates do not fit your use case, you could start off with running `llama stack build` which will allow you to a interactively enter wizard where you will be prompted to enter build configurations. + +It would be best to start with a template and understand the structure of the config file and the various concepts ( APIS, providers, resources, etc.) before starting from scratch. +``` +llama stack build + +> Enter a name for your Llama Stack (e.g. my-local-stack): my-stack +> Enter the image type you want your Llama Stack to be built as (container or venv): venv + +Llama Stack is composed of several APIs working together. Let's select +the provider types (implementations) you want to use for these APIs. + +Tip: use to see options for the providers. + +> Enter provider for API inference: inline::meta-reference +> Enter provider for API safety: inline::llama-guard +> Enter provider for API agents: inline::meta-reference +> Enter provider for API memory: inline::faiss +> Enter provider for API datasetio: inline::meta-reference +> Enter provider for API scoring: inline::meta-reference +> Enter provider for API eval: inline::meta-reference +> Enter provider for API telemetry: inline::meta-reference + + > (Optional) Enter a short description for your Llama Stack: + +You can now edit ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml and run `llama stack run ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml` +``` + + +- In addition to templates, you may customize the build to your liking through editing config files and build from config files with the following command. + +- The config file will be of contents like the ones in `llama_stack/distributions/*build.yaml`. + +``` +llama stack build --config llama_stack/distributions/starter/build.yaml +``` + + + +Llama Stack supports external providers that live outside of the main codebase. This allows you to create and maintain your own providers independently or use community-provided providers. + +To build a distribution with external providers, you need to: + +1. Configure the `external_providers_dir` in your build configuration file: + +```yaml +# Example my-external-stack.yaml with external providers +version: '2' +distribution_spec: + description: Custom distro for CI tests + providers: + inference: + - remote::custom_ollama +# Add more providers as needed +image_type: container +image_name: ci-test +# Path to external provider implementations +external_providers_dir: ~/.llama/providers.d +``` + +Here's an example for a custom Ollama provider: + +```yaml +adapter: + adapter_type: custom_ollama + pip_packages: + - ollama + - aiohttp + - llama-stack-provider-ollama # This is the provider package + config_class: llama_stack_ollama_provider.config.OllamaImplConfig + module: llama_stack_ollama_provider +api_dependencies: [] +optional_api_dependencies: [] +``` + +The `pip_packages` section lists the Python packages required by the provider, as well as the +provider package itself. The package must be available on PyPI or can be provided from a local +directory or a git repository (git must be installed on the build environment). + +2. Build your distribution using the config file: + +``` +llama stack build --config my-external-stack.yaml +``` + +For more information on external providers, including directory structure, provider types, and implementation requirements, see the [External Providers documentation](../providers/external/). + + + +:::tip Podman Alternative +Podman is supported as an alternative to Docker. Set `CONTAINER_BINARY` to `podman` in your environment to use Podman. +::: + +To build a container image, you may start off from a template and use the `--image-type container` flag to specify `container` as the build image type. + +``` +llama stack build --distro starter --image-type container +``` + +``` +$ llama stack build --distro starter --image-type container +... +Containerfile created successfully in /tmp/tmp.viA3a3Rdsg/ContainerfileFROM python:3.10-slim +... +``` + +You can now edit ~/meta-llama/llama-stack/tmp/configs/ollama-run.yaml and run `llama stack run ~/meta-llama/llama-stack/tmp/configs/ollama-run.yaml` +``` + +Now set some environment variables for the inference model ID and Llama Stack Port and create a local directory to mount into the container's file system. + +```bash +export INFERENCE_MODEL="llama3.2:3b" +export LLAMA_STACK_PORT=8321 +mkdir -p ~/.llama +``` + +After this step is successful, you should be able to find the built container image and test it with the below Docker command: + +``` +docker run -d \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ~/.llama:/root/.llama \ + localhost/distribution-ollama:dev \ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=$INFERENCE_MODEL \ + --env OLLAMA_URL=http://host.docker.internal:11434 +``` + +Here are the docker flags and their uses: + +* `-d`: Runs the container in the detached mode as a background process + +* `-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT`: Maps the container port to the host port for accessing the server + +* `-v ~/.llama:/root/.llama`: Mounts the local .llama directory to persist configurations and data + +* `localhost/distribution-ollama:dev`: The name and tag of the container image to run + +* `--port $LLAMA_STACK_PORT`: Port number for the server to listen on + +* `--env INFERENCE_MODEL=$INFERENCE_MODEL`: Sets the model to use for inference + +* `--env OLLAMA_URL=http://host.docker.internal:11434`: Configures the URL for the Ollama service + + + + + +### Running your Stack server +Now, let's start the Llama Stack Distribution Server. You will need the YAML configuration file which was written out at the end by the `llama stack build` step. + +``` +llama stack run -h +usage: llama stack run [-h] [--port PORT] [--image-name IMAGE_NAME] [--env KEY=VALUE] + [--image-type {venv}] [--enable-ui] + [config | template] + +Start the server for a Llama Stack Distribution. You should have already built (or downloaded) and configured the distribution. + +positional arguments: + config | template Path to config file to use for the run or name of known template (`llama stack list` for a list). (default: None) + +options: + -h, --help show this help message and exit + --port PORT Port to run the server on. It can also be passed via the env var LLAMA_STACK_PORT. (default: 8321) + --image-name IMAGE_NAME + Name of the image to run. Defaults to the current environment (default: None) + --env KEY=VALUE Environment variables to pass to the server in KEY=VALUE format. Can be specified multiple times. (default: None) + --image-type {venv} + Image Type used during the build. This should be venv. (default: None) + --enable-ui Start the UI server (default: False) +``` + +**Note:** Container images built with `llama stack build --image-type container` cannot be run using `llama stack run`. Instead, they must be run directly using Docker or Podman commands as shown in the container building section above. + +``` +# Start using template name +llama stack run tgi + +# Start using config file +llama stack run ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml + +# Start using a venv +llama stack run --image-type venv ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml +``` + +``` +$ llama stack run ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml + +Serving API inspect + GET /health + GET /providers/list + GET /routes/list +Serving API inference + POST /inference/chat_completion + POST /inference/completion + POST /inference/embeddings +... +Serving API agents + POST /agents/create + POST /agents/session/create + POST /agents/turn/create + POST /agents/delete + POST /agents/session/delete + POST /agents/session/get + POST /agents/step/get + POST /agents/turn/get + +Listening on ['::', '0.0.0.0']:8321 +INFO: Started server process [2935911] +INFO: Waiting for application startup. +INFO: Application startup complete. +INFO: Uvicorn running on http://['::', '0.0.0.0']:8321 (Press CTRL+C to quit) +INFO: 2401:db00:35c:2d2b:face:0:c9:0:54678 - "GET /models/list HTTP/1.1" 200 OK +``` + +### Listing Distributions +Using the list command, you can view all existing Llama Stack distributions, including stacks built from templates, from scratch, or using custom configuration files. + +``` +llama stack list -h +usage: llama stack list [-h] + +list the build stacks + +options: + -h, --help show this help message and exit +``` + +Example Usage + +``` +llama stack list +``` + +``` +------------------------------+-----------------------------------------------------------------+--------------+------------+ +| Stack Name | Path | Build Config | Run Config | ++------------------------------+-----------------------------------------------------------------------------+--------------+ +| together | ~/.llama/distributions/together | Yes | No | ++------------------------------+-----------------------------------------------------------------------------+--------------+ +| bedrock | ~/.llama/distributions/bedrock | Yes | No | ++------------------------------+-----------------------------------------------------------------------------+--------------+ +| starter | ~/.llama/distributions/starter | Yes | Yes | ++------------------------------+-----------------------------------------------------------------------------+--------------+ +| remote-vllm | ~/.llama/distributions/remote-vllm | Yes | Yes | ++------------------------------+-----------------------------------------------------------------------------+--------------+ +``` + +### Removing a Distribution +Use the remove command to delete a distribution you've previously built. + +``` +llama stack rm -h +usage: llama stack rm [-h] [--all] [name] + +Remove the build stack + +positional arguments: + name Name of the stack to delete (default: None) + +options: + -h, --help show this help message and exit + --all, -a Delete all stacks (use with caution) (default: False) +``` + +Example +``` +llama stack rm llamastack-test +``` + +To keep your environment organized and avoid clutter, consider using `llama stack list` to review old or unused distributions and `llama stack rm ` to delete them when they're no longer needed. + +### Troubleshooting + +If you encounter any issues, ask questions in our discord or search through our [GitHub Issues](https://github.com/meta-llama/llama-stack/issues), or file an new issue. diff --git a/versioned_docs/version-v0.2.23/distributions/configuration.mdx b/versioned_docs/version-v0.2.23/distributions/configuration.mdx new file mode 100644 index 0000000..d87c7f6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/configuration.mdx @@ -0,0 +1,808 @@ +--- +title: Configuring a "Stack" +description: Configuring a "Stack" +sidebar_label: Configuring a "Stack" +sidebar_position: 6 +--- +# Configuring a "Stack" + +The Llama Stack runtime configuration is specified as a YAML file. Here is a simplified version of an example configuration file for the Ollama distribution: + +```{note} +The default `run.yaml` files generated by templates are starting points for your configuration. For guidance on customizing these files for your specific needs, see [Customizing Your run.yaml Configuration](customizing_run_yaml.md). +``` + +```{dropdown} ๐Ÿ‘‹ Click here for a Sample Configuration File + +```yaml +version: 2 +apis: +- agents +- inference +- vector_io +- safety +- telemetry +providers: + inference: + - provider_id: ollama + provider_type: remote::ollama + config: + url: ${env.OLLAMA_URL:=http://localhost:11434} + vector_io: + - provider_id: faiss + provider_type: inline::faiss + config: + kvstore: + type: sqlite + namespace: null + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/distributions/ollama}/faiss_store.db + safety: + - provider_id: llama-guard + provider_type: inline::llama-guard + config: {} + agents: + - provider_id: meta-reference + provider_type: inline::meta-reference + config: + persistence_store: + type: sqlite + namespace: null + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/distributions/ollama}/agents_store.db + telemetry: + - provider_id: meta-reference + provider_type: inline::meta-reference + config: {} +metadata_store: + namespace: null + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/distributions/ollama}/registry.db +models: +- metadata: {} + model_id: ${env.INFERENCE_MODEL} + provider_id: ollama + provider_model_id: null +shields: [] +server: + port: 8321 + auth: + provider_config: + type: "oauth2_token" + jwks: + uri: "https://my-token-issuing-svc.com/jwks" +``` + +Let's break this down into the different sections. The first section specifies the set of APIs that the stack server will serve: +```yaml +apis: +- agents +- inference +- vector_io +- safety +- telemetry +``` + +## Providers +Next up is the most critical part: the set of providers that the stack will use to serve the above APIs. Consider the `inference` API: +```yaml +providers: + inference: + # provider_id is a string you can choose freely + - provider_id: ollama + # provider_type is a string that specifies the type of provider. + # in this case, the provider for inference is ollama and it runs remotely (outside of the distribution) + provider_type: remote::ollama + # config is a dictionary that contains the configuration for the provider. + # in this case, the configuration is the url of the ollama server + config: + url: ${env.OLLAMA_URL:=http://localhost:11434} +``` +A few things to note: +- A _provider instance_ is identified with an (id, type, config) triplet. +- The id is a string you can choose freely. +- You can instantiate any number of provider instances of the same type. +- The configuration dictionary is provider-specific. +- Notice that configuration can reference environment variables (with default values), which are expanded at runtime. When you run a stack server (via docker or via `llama stack run`), you can specify `--env OLLAMA_URL=http://my-server:11434` to override the default value. + +### Environment Variable Substitution + +Llama Stack supports environment variable substitution in configuration values using the +`${env.VARIABLE_NAME}` syntax. This allows you to externalize configuration values and provide +different settings for different environments. The syntax is inspired by [bash parameter expansion](https://www.gnu.org/software/bash/manual/html_node/Shell-Parameter-Expansion.html) +and follows similar patterns. + +#### Basic Syntax + +The basic syntax for environment variable substitution is: + +```yaml +config: + api_key: ${env.API_KEY} + url: ${env.SERVICE_URL} +``` + +If the environment variable is not set, the server will raise an error during startup. + +#### Default Values + +You can provide default values using the `:=` operator: + +```yaml +config: + url: ${env.OLLAMA_URL:=http://localhost:11434} + port: ${env.PORT:=8321} + timeout: ${env.TIMEOUT:=60} +``` + +If the environment variable is not set, the default value `http://localhost:11434` will be used. +Empty defaults are allowed so `url: ${env.OLLAMA_URL:=}` will be set to `None` if the environment variable is not set. + +#### Conditional Values + +You can use the `:+` operator to provide a value only when the environment variable is set: + +```yaml +config: + # Only include this field if ENVIRONMENT is set + environment: ${env.ENVIRONMENT:+production} +``` + +If the environment variable is set, the value after `:+` will be used. If it's not set, the field +will be omitted with a `None` value. + +Do not use conditional values (`${env.OLLAMA_URL:+}`) for empty defaults (`${env.OLLAMA_URL:=}`). +This will be set to `None` if the environment variable is not set. +Conditional must only be used when the environment variable is set. + +#### Examples + +Here are some common patterns: + +```yaml +# Required environment variable (will error if not set) +api_key: ${env.OPENAI_API_KEY} + +# Optional with default +base_url: ${env.API_BASE_URL:=https://api.openai.com/v1} + +# Conditional field +debug_mode: ${env.DEBUG:+true} + +# Optional field that becomes None if not set +optional_token: ${env.OPTIONAL_TOKEN:+} +``` + +#### Runtime Override + +You can override environment variables at runtime when starting the server: + +```bash +# Override specific environment variables +llama stack run --config run.yaml --env API_KEY=sk-123 --env BASE_URL=https://custom-api.com + +# Or set them in your shell +export API_KEY=sk-123 +export BASE_URL=https://custom-api.com +llama stack run --config run.yaml +``` + +#### Type Safety + +The environment variable substitution system is type-safe: + +- String values remain strings +- Empty defaults (`${env.VAR:+}`) are converted to `None` for fields that accept `str | None` +- Numeric defaults are properly typed (e.g., `${env.PORT:=8321}` becomes an integer) +- Boolean defaults work correctly (e.g., `${env.DEBUG:=false}` becomes a boolean) + +## Resources + +Let's look at the `models` section: + +```yaml +models: +- metadata: {} + model_id: ${env.INFERENCE_MODEL} + provider_id: ollama + provider_model_id: null + model_type: llm +``` +A Model is an instance of a "Resource" (see [Concepts](../concepts/)) and is associated with a specific inference provider (in this case, the provider with identifier `ollama`). This is an instance of a "pre-registered" model. While we always encourage the clients to register models before using them, some Stack servers may come up a list of "already known and available" models. + +What's with the `provider_model_id` field? This is an identifier for the model inside the provider's model catalog. Contrast it with `model_id` which is the identifier for the same model for Llama Stack's purposes. For example, you may want to name "llama3.2:vision-11b" as "image_captioning_model" when you use it in your Stack interactions. When omitted, the server will set `provider_model_id` to be the same as `model_id`. + +If you need to conditionally register a model in the configuration, such as only when specific environment variable(s) are set, this can be accomplished by utilizing a special `__disabled__` string as the default value of an environment variable substitution, as shown below: + +```yaml +models: +- metadata: {} + model_id: ${env.INFERENCE_MODEL:__disabled__} + provider_id: ollama + provider_model_id: ${env.INFERENCE_MODEL:__disabled__} +``` + +The snippet above will only register this model if the environment variable `INFERENCE_MODEL` is set and non-empty. If the environment variable is not set, the model will not get registered at all. + +## Server Configuration + +The `server` section configures the HTTP server that serves the Llama Stack APIs: + +```yaml +server: + port: 8321 # Port to listen on (default: 8321) + tls_certfile: "/path/to/cert.pem" # Optional: Path to TLS certificate for HTTPS + tls_keyfile: "/path/to/key.pem" # Optional: Path to TLS key for HTTPS + cors: true # Optional: Enable CORS (dev mode) or full config object +``` + +### CORS Configuration + +CORS (Cross-Origin Resource Sharing) can be configured in two ways: + +**Local development** (allows localhost origins only): +```yaml +server: + cors: true +``` + +**Explicit configuration** (custom origins and settings): +```yaml +server: + cors: + allow_origins: ["https://myapp.com", "https://app.example.com"] + allow_methods: ["GET", "POST", "PUT", "DELETE"] + allow_headers: ["Content-Type", "Authorization"] + allow_credentials: true + max_age: 3600 +``` + +When `cors: true`, the server enables secure localhost-only access for local development. For production, specify exact origins to maintain security. + +### Authentication Configuration + +> **Breaking Change (v0.2.14)**: The authentication configuration structure has changed. The previous format with `provider_type` and `config` fields has been replaced with a unified `provider_config` field that includes the `type` field. Update your configuration files accordingly. + +The `auth` section configures authentication for the server. When configured, all API requests must include a valid Bearer token in the Authorization header: + +``` +Authorization: Bearer +``` + +The server supports multiple authentication providers: + +#### OAuth 2.0/OpenID Connect Provider with Kubernetes + +The server can be configured to use service account tokens for authorization, validating these against the Kubernetes API server, e.g.: +```yaml +server: + auth: + provider_config: + type: "oauth2_token" + jwks: + uri: "https://kubernetes.default.svc:8443/openid/v1/jwks" + token: "${env.TOKEN:+}" + key_recheck_period: 3600 + tls_cafile: "/path/to/ca.crt" + issuer: "https://kubernetes.default.svc" + audience: "https://kubernetes.default.svc" +``` + +To find your cluster's jwks uri (from which the public key(s) to verify the token signature are obtained), run: +``` +kubectl get --raw /.well-known/openid-configuration| jq -r .jwks_uri +``` + +For the tls_cafile, you can use the CA certificate of the OIDC provider: +```bash +kubectl config view --minify -o jsonpath='{.clusters[0].cluster.certificate-authority}' +``` + +For the issuer, you can use the OIDC provider's URL: +```bash +kubectl get --raw /.well-known/openid-configuration| jq .issuer +``` + +The audience can be obtained from a token, e.g. run: +```bash +kubectl create token default --duration=1h | cut -d. -f2 | base64 -d | jq .aud +``` + +The jwks token is used to authorize access to the jwks endpoint. You can obtain a token by running: + +```bash +kubectl create namespace llama-stack +kubectl create serviceaccount llama-stack-auth -n llama-stack +kubectl create token llama-stack-auth -n llama-stack > llama-stack-auth-token +export TOKEN=$(cat llama-stack-auth-token) +``` + +Alternatively, you can configure the jwks endpoint to allow anonymous access. To do this, make sure +the `kube-apiserver` runs with `--anonymous-auth=true` to allow unauthenticated requests +and that the correct RoleBinding is created to allow the service account to access the necessary +resources. If that is not the case, you can create a RoleBinding for the service account to access +the necessary resources: + +```yaml +# allow-anonymous-openid.yaml +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRole +metadata: + name: allow-anonymous-openid +rules: +- nonResourceURLs: ["/openid/v1/jwks"] + verbs: ["get"] +--- +apiVersion: rbac.authorization.k8s.io/v1 +kind: ClusterRoleBinding +metadata: + name: allow-anonymous-openid +roleRef: + apiGroup: rbac.authorization.k8s.io + kind: ClusterRole + name: allow-anonymous-openid +subjects: +- kind: User + name: system:anonymous + apiGroup: rbac.authorization.k8s.io +``` + +And then apply the configuration: +```bash +kubectl apply -f allow-anonymous-openid.yaml +``` + +The provider extracts user information from the JWT token: +- Username from the `sub` claim becomes a role +- Kubernetes groups become teams + +You can easily validate a request by running: + +```bash +curl -s -L -H "Authorization: Bearer $(cat llama-stack-auth-token)" http://127.0.0.1:8321/v1/providers +``` + +#### Kubernetes Authentication Provider + +The server can be configured to use Kubernetes SelfSubjectReview API to validate tokens directly against the Kubernetes API server: + +```yaml +server: + auth: + provider_config: + type: "kubernetes" + api_server_url: "https://kubernetes.default.svc" + claims_mapping: + username: "roles" + groups: "roles" + uid: "uid_attr" + verify_tls: true + tls_cafile: "/path/to/ca.crt" +``` + +Configuration options: +- `api_server_url`: The Kubernetes API server URL (e.g., https://kubernetes.default.svc:6443) +- `verify_tls`: Whether to verify TLS certificates (default: true) +- `tls_cafile`: Path to CA certificate file for TLS verification +- `claims_mapping`: Mapping of Kubernetes user claims to access attributes + +The provider validates tokens by sending a SelfSubjectReview request to the Kubernetes API server at `/apis/authentication.k8s.io/v1/selfsubjectreviews`. The provider extracts user information from the response: +- Username from the `userInfo.username` field +- Groups from the `userInfo.groups` field +- UID from the `userInfo.uid` field + +To obtain a token for testing: +```bash +kubectl create namespace llama-stack +kubectl create serviceaccount llama-stack-auth -n llama-stack +kubectl create token llama-stack-auth -n llama-stack > llama-stack-auth-token +``` + +You can validate a request by running: +```bash +curl -s -L -H "Authorization: Bearer $(cat llama-stack-auth-token)" http://127.0.0.1:8321/v1/providers +``` + +#### GitHub Token Provider +Validates GitHub personal access tokens or OAuth tokens directly: +```yaml +server: + auth: + provider_config: + type: "github_token" + github_api_base_url: "https://api.github.com" # Or GitHub Enterprise URL +``` + +The provider fetches user information from GitHub and maps it to access attributes based on the `claims_mapping` configuration. + +#### Custom Provider +Validates tokens against a custom authentication endpoint: +```yaml +server: + auth: + provider_config: + type: "custom" + endpoint: "https://auth.example.com/validate" # URL of the auth endpoint +``` + +The custom endpoint receives a POST request with: +```json +{ + "api_key": "", + "request": { + "path": "/api/v1/endpoint", + "headers": { + "content-type": "application/json", + "user-agent": "curl/7.64.1" + }, + "params": { + "key": ["value"] + } + } +} +``` + +And must respond with: +```json +{ + "access_attributes": { + "roles": ["admin", "user"], + "teams": ["ml-team", "nlp-team"], + "projects": ["llama-3", "project-x"], + "namespaces": ["research"] + }, + "message": "Authentication successful" +} +``` + +If no access attributes are returned, the token is used as a namespace. + +### Access control + +When authentication is enabled, access to resources is controlled +through the `access_policy` attribute of the auth config section under +server. The value for this is a list of access rules. + +Each access rule defines a list of actions either to permit or to +forbid. It may specify a principal or a resource that must match for +the rule to take effect. + +Valid actions are create, read, update, and delete. The resource to +match should be specified in the form of a type qualified identifier, +e.g. model::my-model or vector_db::some-db, or a wildcard for all +resources of a type, e.g. model::*. If the principal or resource are +not specified, they will match all requests. + +The valid resource types are model, shield, vector_db, dataset, +scoring_function, benchmark, tool, tool_group and session. + +A rule may also specify a condition, either a 'when' or an 'unless', +with additional constraints as to where the rule applies. The +constraints supported at present are: + + - 'user with `` in ``' + - 'user with `` not in ``' + - 'user is owner' + - 'user is not owner' + - 'user in owners ``' + - 'user not in owners ``' + +The attributes defined for a user will depend on how the auth +configuration is defined. + +When checking whether a particular action is allowed by the current +user for a resource, all the defined rules are tested in order to find +a match. If a match is found, the request is permitted or forbidden +depending on the type of rule. If no match is found, the request is +denied. + +If no explicit rules are specified, a default policy is defined with +which all users can access all resources defined in config but +resources created dynamically can only be accessed by the user that +created them. + +Examples: + +The following restricts access to particular github users: + +```yaml +server: + auth: + provider_config: + type: "github_token" + github_api_base_url: "https://api.github.com" + access_policy: + - permit: + principal: user-1 + actions: [create, read, delete] + description: user-1 has full access to all resources + - permit: + principal: user-2 + actions: [read] + resource: model::model-1 + description: user-2 has read access to model-1 only +``` + +Similarly, the following restricts access to particular kubernetes +service accounts: + +```yaml +server: + auth: + provider_config: + type: "oauth2_token" + audience: https://kubernetes.default.svc.cluster.local + issuer: https://kubernetes.default.svc.cluster.local + tls_cafile: /home/gsim/.minikube/ca.crt + jwks: + uri: https://kubernetes.default.svc.cluster.local:8443/openid/v1/jwks + token: ${env.TOKEN} + access_policy: + - permit: + principal: system:serviceaccount:my-namespace:my-serviceaccount + actions: [create, read, delete] + description: specific serviceaccount has full access to all resources + - permit: + principal: system:serviceaccount:default:default + actions: [read] + resource: model::model-1 + description: default account has read access to model-1 only +``` + +The following policy, which assumes that users are defined with roles +and teams by whichever authentication system is in use, allows any +user with a valid token to use models, create resources other than +models, read and delete resources they created and read resources +created by users sharing a team with them: + +``` + access_policy: + - permit: + actions: [read] + resource: model::* + description: all users have read access to models + - forbid: + actions: [create, delete] + resource: model::* + unless: user with admin in roles + description: only user with admin role can create or delete models + - permit: + actions: [create, read, delete] + when: user is owner + description: users can create resources other than models and read and delete those they own + - permit: + actions: [read] + when: user in owner teams + description: any user has read access to any resource created by a user with the same team +``` + +#### API Endpoint Authorization with Scopes + +In addition to resource-based access control, Llama Stack supports endpoint-level authorization using OAuth 2.0 style scopes. When authentication is enabled, specific API endpoints require users to have particular scopes in their authentication token. + +**Scope-Gated APIs:** +The following APIs are currently gated by scopes: + +- **Telemetry API** (scope: `telemetry.read`): + - `POST /telemetry/traces` - Query traces + - `GET /telemetry/traces/{trace_id}` - Get trace by ID + - `GET /telemetry/traces/{trace_id}/spans/{span_id}` - Get span by ID + - `POST /telemetry/spans/{span_id}/tree` - Get span tree + - `POST /telemetry/spans` - Query spans + - `POST /telemetry/metrics/{metric_name}` - Query metrics + +**Authentication Configuration:** + +For **JWT/OAuth2 providers**, scopes should be included in the JWT's claims: +```json +{ + "sub": "user123", + "scope": "telemetry.read", + "aud": "llama-stack" +} +``` + +For **custom authentication providers**, the endpoint must return user attributes including the `scopes` array: +```json +{ + "principal": "user123", + "attributes": { + "scopes": ["telemetry.read"] + } +} +``` + +**Behavior:** +- Users without the required scope receive a 403 Forbidden response +- When authentication is disabled, scope checks are bypassed +- Endpoints without `required_scope` work normally for all authenticated users + +### Quota Configuration + +The `quota` section allows you to enable server-side request throttling for both +authenticated and anonymous clients. This is useful for preventing abuse, enforcing +fairness across tenants, and controlling infrastructure costs without requiring +client-side rate limiting or external proxies. + +Quotas are disabled by default. When enabled, each client is tracked using either: + +* Their authenticated `client_id` (derived from the Bearer token), or +* Their IP address (fallback for anonymous requests) + +Quota state is stored in a SQLite-backed key-value store, and rate limits are applied +within a configurable time window (currently only `day` is supported). + +#### Example + +```yaml +server: + quota: + kvstore: + type: sqlite + db_path: ./quotas.db + anonymous_max_requests: 100 + authenticated_max_requests: 1000 + period: day +``` + +#### Configuration Options + +| Field | Description | +| ---------------------------- | -------------------------------------------------------------------------- | +| `kvstore` | Required. Backend storage config for tracking request counts. | +| `kvstore.type` | Must be `"sqlite"` for now. Other backends may be supported in the future. | +| `kvstore.db_path` | File path to the SQLite database. | +| `anonymous_max_requests` | Max requests per period for unauthenticated clients. | +| `authenticated_max_requests` | Max requests per period for authenticated clients. | +| `period` | Time window for quota enforcement. Only `"day"` is supported. | + +> Note: if `authenticated_max_requests` is set but no authentication provider is +configured, the server will fall back to applying `anonymous_max_requests` to all +clients. + +#### Example with Authentication Enabled + +```yaml +server: + port: 8321 + auth: + provider_config: + type: custom + endpoint: https://auth.example.com/validate + quota: + kvstore: + type: sqlite + db_path: ./quotas.db + anonymous_max_requests: 100 + authenticated_max_requests: 1000 + period: day +``` + +If a client exceeds their limit, the server responds with: + +```http +HTTP/1.1 429 Too Many Requests +Content-Type: application/json + +{ + "error": { + "message": "Quota exceeded" + } +} +``` + +### CORS Configuration + +Configure CORS to allow web browsers to make requests from different domains. Disabled by default. + +#### Quick Setup + +For development, use the simple boolean flag: + +```yaml +server: + cors: true # Auto-enables localhost with any port +``` + +This automatically allows `http://localhost:*` and `https://localhost:*` with secure defaults. + +#### Custom Configuration + +For specific origins and full control: + +```yaml +server: + cors: + allow_origins: ["https://myapp.com", "https://staging.myapp.com"] + allow_credentials: true + allow_methods: ["GET", "POST", "PUT", "DELETE"] + allow_headers: ["Content-Type", "Authorization"] + allow_origin_regex: "https://.*\\.example\\.com" # Optional regex pattern + expose_headers: ["X-Total-Count"] + max_age: 86400 +``` + +#### Configuration Options + +| Field | Description | Default | +| -------------------- | ---------------------------------------------- | ------- | +| `allow_origins` | List of allowed origins. Use `["*"]` for any. | `["*"]` | +| `allow_origin_regex` | Regex pattern for allowed origins (optional). | `None` | +| `allow_methods` | Allowed HTTP methods. | `["*"]` | +| `allow_headers` | Allowed headers. | `["*"]` | +| `allow_credentials` | Allow credentials (cookies, auth headers). | `false` | +| `expose_headers` | Headers exposed to browser. | `[]` | +| `max_age` | Preflight cache time (seconds). | `600` | + +**Security Notes**: +- `allow_credentials: true` requires explicit origins (no wildcards) +- `cors: true` enables localhost access only (secure for development) +- For public APIs, always specify exact allowed origins + +## Extending to handle Safety + +Configuring Safety can be a little involved so it is instructive to go through an example. + +The Safety API works with the associated Resource called a `Shield`. Providers can support various kinds of Shields. Good examples include the [Llama Guard](https://ai.meta.com/research/publications/llama-guard-llm-based-input-output-safeguard-for-human-ai-conversations/) system-safety models, or [Bedrock Guardrails](https://aws.amazon.com/bedrock/guardrails/). + +To configure a Bedrock Shield, you would need to add: +- A Safety API provider instance with type `remote::bedrock` +- A Shield resource served by this provider. + +```yaml +... +providers: + safety: + - provider_id: bedrock + provider_type: remote::bedrock + config: + aws_access_key_id: ${env.AWS_ACCESS_KEY_ID} + aws_secret_access_key: ${env.AWS_SECRET_ACCESS_KEY} +... +shields: +- provider_id: bedrock + params: + guardrailVersion: ${env.GUARDRAIL_VERSION} + provider_shield_id: ${env.GUARDRAIL_ID} +... +``` + +The situation is more involved if the Shield needs _Inference_ of an associated model. This is the case with Llama Guard. In that case, you would need to add: +- A Safety API provider instance with type `inline::llama-guard` +- An Inference API provider instance for serving the model. +- A Model resource associated with this provider. +- A Shield resource served by the Safety provider. + +The yaml configuration for this setup, assuming you were using vLLM as your inference server, would look like: +```yaml +... +providers: + safety: + - provider_id: llama-guard + provider_type: inline::llama-guard + config: {} + inference: + # this vLLM server serves the "normal" inference model (e.g., llama3.2:3b) + - provider_id: vllm-0 + provider_type: remote::vllm + config: + url: ${env.VLLM_URL:=http://localhost:8000} + # this vLLM server serves the llama-guard model (e.g., llama-guard:3b) + - provider_id: vllm-1 + provider_type: remote::vllm + config: + url: ${env.SAFETY_VLLM_URL:=http://localhost:8001} +... +models: +- metadata: {} + model_id: ${env.INFERENCE_MODEL} + provider_id: vllm-0 + provider_model_id: null +- metadata: {} + model_id: ${env.SAFETY_MODEL} + provider_id: vllm-1 + provider_model_id: null +shields: +- provider_id: llama-guard + shield_id: ${env.SAFETY_MODEL} # Llama Guard shields are identified by the corresponding LlamaGuard model + provider_shield_id: null +... +``` diff --git a/versioned_docs/version-v0.2.23/distributions/customizing_run_yaml.mdx b/versioned_docs/version-v0.2.23/distributions/customizing_run_yaml.mdx new file mode 100644 index 0000000..513712f --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/customizing_run_yaml.mdx @@ -0,0 +1,46 @@ +--- +title: Customizing run.yaml +description: Customizing run.yaml files for Llama Stack templates +sidebar_label: Customizing run.yaml +sidebar_position: 4 +--- +# Customizing run.yaml Files + +The `run.yaml` files generated by Llama Stack templates are **starting points** designed to be customized for your specific needs. They are not meant to be used as-is in production environments. + +## Key Points + +- **Templates are starting points**: Generated `run.yaml` files contain defaults for development/testing +- **Customization expected**: Update URLs, credentials, models, and settings for your environment +- **Version control separately**: Keep customized configs in your own repository +- **Environment-specific**: Create different configurations for dev, staging, production + +## What You Can Customize + +You can customize: +- **Provider endpoints**: Change `http://localhost:8000` to your actual servers +- **Swap providers**: Replace default providers (e.g., swap Tavily with Brave for search) +- **Storage paths**: Move from `/tmp/` to production directories +- **Authentication**: Add API keys, SSL, timeouts +- **Models**: Different model sizes for dev vs prod +- **Database settings**: Switch from SQLite to PostgreSQL +- **Tool configurations**: Add custom tools and integrations + +## Best Practices + +- Use environment variables for secrets and environment-specific values +- Create separate `run.yaml` files for different environments (dev, staging, prod) +- Document your changes with comments +- Test configurations before deployment +- Keep your customized configs in version control + +Example structure: +``` +your-project/ +โ”œโ”€โ”€ configs/ +โ”‚ โ”œโ”€โ”€ dev-run.yaml +โ”‚ โ”œโ”€โ”€ prod-run.yaml +โ””โ”€โ”€ README.md +``` + +The goal is to take the generated template and adapt it to your specific infrastructure and operational needs. diff --git a/versioned_docs/version-v0.2.23/distributions/eks/apply.sh b/versioned_docs/version-v0.2.23/distributions/eks/apply.sh new file mode 100755 index 0000000..3ad3dd2 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/eks/apply.sh @@ -0,0 +1,19 @@ +#!/usr/bin/env bash + +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +set -euo pipefail + +SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" +K8S_DIR="${SCRIPT_DIR}/../k8s" + +echo "Setting up AWS EKS-specific storage class..." +kubectl apply -f gp3-topology-aware.yaml + +echo "Running main Kubernetes deployment..." +cd "${K8S_DIR}" +./apply.sh "$@" diff --git a/versioned_docs/version-v0.2.23/distributions/eks/gp3-topology-aware.yaml b/versioned_docs/version-v0.2.23/distributions/eks/gp3-topology-aware.yaml new file mode 100644 index 0000000..1192ba1 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/eks/gp3-topology-aware.yaml @@ -0,0 +1,15 @@ +# Set up default storage class on AWS EKS +apiVersion: storage.k8s.io/v1 +kind: StorageClass +metadata: + name: gp3-topology-aware + annotations: + storageclass.kubernetes.io/is-default-class: "true" +parameters: + type: gp3 + iops: "3000" + throughput: "125" +provisioner: ebs.csi.aws.com +reclaimPolicy: Delete +volumeBindingMode: WaitForFirstConsumer +allowVolumeExpansion: true diff --git a/versioned_docs/version-v0.2.23/distributions/importing_as_library.mdx b/versioned_docs/version-v0.2.23/distributions/importing_as_library.mdx new file mode 100644 index 0000000..122e522 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/importing_as_library.mdx @@ -0,0 +1,40 @@ +--- +title: Using Llama Stack as a Library +description: How to use Llama Stack as a Python library instead of running a server +sidebar_label: Importing as Library +sidebar_position: 5 +--- +# Using Llama Stack as a Library + +## Setup Llama Stack without a Server +If you are planning to use an external service for Inference (even Ollama or TGI counts as external), it is often easier to use Llama Stack as a library. +This avoids the overhead of setting up a server. +```bash +# setup +uv pip install llama-stack +llama stack build --distro starter --image-type venv +``` + +```python +from llama_stack.core.library_client import LlamaStackAsLibraryClient + +client = LlamaStackAsLibraryClient( + "starter", + # provider_data is optional, but if you need to pass in any provider specific data, you can do so here. + provider_data={"tavily_search_api_key": os.environ["TAVILY_SEARCH_API_KEY"]}, +) +``` + +This will parse your config and set up any inline implementations and remote clients needed for your implementation. + +Then, you can access the APIs like `models` and `inference` on the client and call their methods directly: + +```python +response = client.models.list() +``` + +If you've created a [custom distribution](./building_distro), you can also use the run.yaml configuration file directly: + +```python +client = LlamaStackAsLibraryClient(config_path) +``` diff --git a/versioned_docs/version-v0.2.23/distributions/index.mdx b/versioned_docs/version-v0.2.23/distributions/index.mdx new file mode 100644 index 0000000..0149f14 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/index.mdx @@ -0,0 +1,21 @@ +--- +title: Distributions Overview +description: Pre-packaged sets of Llama Stack components for different deployment scenarios +sidebar_label: Overview +sidebar_position: 1 +--- + +# Distributions Overview + +A distribution is a pre-packaged set of Llama Stack components that can be deployed together. + +This section provides an overview of the distributions available in Llama Stack. + +## Distribution Guides + +- **[Available Distributions](./list_of_distributions.mdx)** - Complete list and comparison of all distributions +- **[Building Custom Distributions](./building_distro.mdx)** - Create your own distribution from scratch +- **[Customizing Configuration](./customizing_run_yaml.mdx)** - Customize run.yaml for your needs +- **[Starting Llama Stack Server](./starting_llama_stack_server.mdx)** - How to run distributions +- **[Importing as Library](./importing_as_library.mdx)** - Use distributions in your code +- **[Configuration Reference](./configuration.mdx)** - Configuration file format details diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/apply.sh b/versioned_docs/version-v0.2.23/distributions/k8s/apply.sh new file mode 100755 index 0000000..1b5b268 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/apply.sh @@ -0,0 +1,63 @@ +#!/usr/bin/env bash + +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +export POSTGRES_USER=llamastack +export POSTGRES_DB=llamastack +export POSTGRES_PASSWORD=llamastack + +export INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct +export SAFETY_MODEL=meta-llama/Llama-Guard-3-1B + +# HF_TOKEN should be set by the user; base64 encode it for the secret +if [ -n "${HF_TOKEN:-}" ]; then + export HF_TOKEN_BASE64=$(echo -n "$HF_TOKEN" | base64) +else + echo "ERROR: HF_TOKEN not set. You need it for vLLM to download models from Hugging Face." + exit 1 +fi + +if [ -z "${GITHUB_CLIENT_ID:-}" ]; then + echo "ERROR: GITHUB_CLIENT_ID not set. You need it for Github login to work. See the Kubernetes Deployment Guide in the Llama Stack documentation." + exit 1 +fi + +if [ -z "${GITHUB_CLIENT_SECRET:-}" ]; then + echo "ERROR: GITHUB_CLIENT_SECRET not set. You need it for Github login to work. See the Kubernetes Deployment Guide in the Llama Stack documentation." + exit 1 +fi + +if [ -z "${LLAMA_STACK_UI_URL:-}" ]; then + echo "ERROR: LLAMA_STACK_UI_URL not set. Should be set to the external URL of the UI (excluding port). You need it for Github login to work. See the Kubernetes Deployment Guide in the Llama Stack documentation." + exit 1 +fi + + + + +set -euo pipefail +set -x + +# Apply the HF token secret if HF_TOKEN is provided +if [ -n "${HF_TOKEN:-}" ]; then + envsubst < ./hf-token-secret.yaml.template | kubectl apply -f - +fi + +envsubst < ./vllm-k8s.yaml.template | kubectl apply -f - +envsubst < ./vllm-safety-k8s.yaml.template | kubectl apply -f - +envsubst < ./postgres-k8s.yaml.template | kubectl apply -f - +envsubst < ./chroma-k8s.yaml.template | kubectl apply -f - + +kubectl create configmap llama-stack-config --from-file=stack_run_config.yaml \ + --dry-run=client -o yaml > stack-configmap.yaml + +kubectl apply -f stack-configmap.yaml + +envsubst < ./stack-k8s.yaml.template | kubectl apply -f - +envsubst < ./ingress-k8s.yaml.template | kubectl apply -f - + +envsubst < ./ui-k8s.yaml.template | kubectl apply -f - diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/chroma-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/chroma-k8s.yaml.template new file mode 100644 index 0000000..a2a5e3b --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/chroma-k8s.yaml.template @@ -0,0 +1,66 @@ +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: chromadb-pvc +spec: + accessModes: + - ReadWriteOnce + resources: + requests: + storage: 20Gi +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: chromadb +spec: + replicas: 1 + selector: + matchLabels: + app: chromadb + template: + metadata: + labels: + app: chromadb + spec: + containers: + - name: chromadb + image: chromadb/chroma:latest + ports: + - containerPort: 6000 + env: + - name: CHROMA_HOST + value: "0.0.0.0" + - name: CHROMA_PORT + value: "6000" + - name: PERSIST_DIRECTORY + value: "/chroma/chroma" + - name: CHROMA_DB_IMPL + value: "duckdb+parquet" + resources: + requests: + memory: "512Mi" + cpu: "250m" + limits: + memory: "2Gi" + cpu: "1000m" + volumeMounts: + - name: chromadb-storage + mountPath: /chroma/chroma + volumes: + - name: chromadb-storage + persistentVolumeClaim: + claimName: chromadb-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: chromadb +spec: + selector: + app: chromadb + ports: + - protocol: TCP + port: 6000 + targetPort: 6000 + type: ClusterIP diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/hf-token-secret.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/hf-token-secret.yaml.template new file mode 100644 index 0000000..b6db8e7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/hf-token-secret.yaml.template @@ -0,0 +1,7 @@ +apiVersion: v1 +kind: Secret +metadata: + name: hf-token-secret +type: Opaque +data: + token: ${HF_TOKEN_BASE64} diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/ingress-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/ingress-k8s.yaml.template new file mode 100644 index 0000000..9ebe86b --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/ingress-k8s.yaml.template @@ -0,0 +1,17 @@ +apiVersion: v1 +kind: Service +metadata: + name: llama-stack-service +spec: + type: LoadBalancer + selector: + app.kubernetes.io/name: llama-stack + ports: + - name: llama-stack-api + port: 8321 + targetPort: 8321 + protocol: TCP + - name: llama-stack-ui + port: 8322 + targetPort: 8322 + protocol: TCP diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/postgres-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/postgres-k8s.yaml.template new file mode 100644 index 0000000..86a7656 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/postgres-k8s.yaml.template @@ -0,0 +1,66 @@ +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: postgres-pvc +spec: + accessModes: + - ReadWriteOnce + resources: + requests: + storage: 10Gi +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: postgres +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: postgres + template: + metadata: + labels: + app.kubernetes.io/name: postgres + spec: + containers: + - name: postgres + image: postgres:15 + env: + - name: POSTGRES_DB + value: "${POSTGRES_DB}" + - name: POSTGRES_USER + value: "${POSTGRES_USER}" + - name: POSTGRES_PASSWORD + value: "${POSTGRES_PASSWORD}" + - name: PGDATA + value: "/var/lib/postgresql/data/pgdata" + ports: + - containerPort: 5432 + resources: + requests: + memory: "512Mi" + cpu: "250m" + limits: + memory: "1Gi" + cpu: "500m" + volumeMounts: + - name: postgres-storage + mountPath: /var/lib/postgresql/data + volumes: + - name: postgres-storage + persistentVolumeClaim: + claimName: postgres-pvc +--- +apiVersion: v1 +kind: Service +metadata: + name: postgres-server +spec: + selector: + app.kubernetes.io/name: postgres + ports: + - protocol: TCP + port: 5432 + targetPort: 5432 + type: ClusterIP diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/stack-configmap.yaml b/versioned_docs/version-v0.2.23/distributions/k8s/stack-configmap.yaml new file mode 100644 index 0000000..3dbb0da --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/stack-configmap.yaml @@ -0,0 +1,56 @@ +apiVersion: v1 +data: + stack_run_config.yaml: "version: '2'\nimage_name: kubernetes-demo\napis:\n- agents\n- + inference\n- files\n- safety\n- telemetry\n- tool_runtime\n- vector_io\nproviders:\n + \ inference:\n - provider_id: vllm-inference\n provider_type: remote::vllm\n + \ config:\n url: ${env.VLLM_URL:=http://localhost:8000/v1}\n max_tokens: + ${env.VLLM_MAX_TOKENS:=4096}\n api_token: ${env.VLLM_API_TOKEN:=fake}\n tls_verify: + ${env.VLLM_TLS_VERIFY:=true}\n - provider_id: vllm-safety\n provider_type: + remote::vllm\n config:\n url: ${env.VLLM_SAFETY_URL:=http://localhost:8000/v1}\n + \ max_tokens: ${env.VLLM_MAX_TOKENS:=4096}\n api_token: ${env.VLLM_API_TOKEN:=fake}\n + \ tls_verify: ${env.VLLM_TLS_VERIFY:=true}\n - provider_id: sentence-transformers\n + \ provider_type: inline::sentence-transformers\n config: {}\n vector_io:\n + \ - provider_id: ${env.ENABLE_CHROMADB:+chromadb}\n provider_type: remote::chromadb\n + \ config:\n url: ${env.CHROMADB_URL:=}\n kvstore:\n type: postgres\n + \ host: ${env.POSTGRES_HOST:=localhost}\n port: ${env.POSTGRES_PORT:=5432}\n + \ db: ${env.POSTGRES_DB:=llamastack}\n user: ${env.POSTGRES_USER:=llamastack}\n + \ password: ${env.POSTGRES_PASSWORD:=llamastack}\n files:\n - provider_id: + meta-reference-files\n provider_type: inline::localfs\n config:\n storage_dir: + ${env.FILES_STORAGE_DIR:=~/.llama/distributions/starter/files}\n metadata_store:\n + \ type: sqlite\n db_path: ${env.SQLITE_STORE_DIR:=~/.llama/distributions/starter}/files_metadata.db + \ \n safety:\n - provider_id: llama-guard\n provider_type: inline::llama-guard\n + \ config:\n excluded_categories: []\n agents:\n - provider_id: meta-reference\n + \ provider_type: inline::meta-reference\n config:\n persistence_store:\n + \ type: postgres\n host: ${env.POSTGRES_HOST:=localhost}\n port: + ${env.POSTGRES_PORT:=5432}\n db: ${env.POSTGRES_DB:=llamastack}\n user: + ${env.POSTGRES_USER:=llamastack}\n password: ${env.POSTGRES_PASSWORD:=llamastack}\n + \ responses_store:\n type: postgres\n host: ${env.POSTGRES_HOST:=localhost}\n + \ port: ${env.POSTGRES_PORT:=5432}\n db: ${env.POSTGRES_DB:=llamastack}\n + \ user: ${env.POSTGRES_USER:=llamastack}\n password: ${env.POSTGRES_PASSWORD:=llamastack}\n + \ telemetry:\n - provider_id: meta-reference\n provider_type: inline::meta-reference\n + \ config:\n service_name: \"${env.OTEL_SERVICE_NAME:=\\u200B}\"\n sinks: + ${env.TELEMETRY_SINKS:=console}\n tool_runtime:\n - provider_id: brave-search\n + \ provider_type: remote::brave-search\n config:\n api_key: ${env.BRAVE_SEARCH_API_KEY:+}\n + \ max_results: 3\n - provider_id: tavily-search\n provider_type: remote::tavily-search\n + \ config:\n api_key: ${env.TAVILY_SEARCH_API_KEY:+}\n max_results: + 3\n - provider_id: rag-runtime\n provider_type: inline::rag-runtime\n config: + {}\n - provider_id: model-context-protocol\n provider_type: remote::model-context-protocol\n + \ config: {}\nmetadata_store:\n type: postgres\n host: ${env.POSTGRES_HOST:=localhost}\n + \ port: ${env.POSTGRES_PORT:=5432}\n db: ${env.POSTGRES_DB:=llamastack}\n user: + ${env.POSTGRES_USER:=llamastack}\n password: ${env.POSTGRES_PASSWORD:=llamastack}\n + \ table_name: llamastack_kvstore\ninference_store:\n type: postgres\n host: + ${env.POSTGRES_HOST:=localhost}\n port: ${env.POSTGRES_PORT:=5432}\n db: ${env.POSTGRES_DB:=llamastack}\n + \ user: ${env.POSTGRES_USER:=llamastack}\n password: ${env.POSTGRES_PASSWORD:=llamastack}\nmodels:\n- + metadata:\n embedding_dimension: 384\n model_id: all-MiniLM-L6-v2\n provider_id: + sentence-transformers\n model_type: embedding\n- metadata: {}\n model_id: ${env.INFERENCE_MODEL}\n + \ provider_id: vllm-inference\n model_type: llm\n- metadata: {}\n model_id: + ${env.SAFETY_MODEL:=meta-llama/Llama-Guard-3-1B}\n provider_id: vllm-safety\n + \ model_type: llm\nshields:\n- shield_id: ${env.SAFETY_MODEL:=meta-llama/Llama-Guard-3-1B}\nvector_dbs: + []\ndatasets: []\nscoring_fns: []\nbenchmarks: []\ntool_groups:\n- toolgroup_id: + builtin::websearch\n provider_id: tavily-search\n- toolgroup_id: builtin::rag\n + \ provider_id: rag-runtime\nserver:\n port: 8321\n auth:\n provider_config:\n + \ type: github_token\n" +kind: ConfigMap +metadata: + creationTimestamp: null + name: llama-stack-config diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/stack-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/stack-k8s.yaml.template new file mode 100644 index 0000000..dfc049f --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/stack-k8s.yaml.template @@ -0,0 +1,69 @@ +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: llama-pvc +spec: + accessModes: + - ReadWriteOnce + resources: + requests: + storage: 1Gi +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: llama-stack-server +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: llama-stack + app.kubernetes.io/component: server + template: + metadata: + labels: + app.kubernetes.io/name: llama-stack + app.kubernetes.io/component: server + spec: + containers: + - name: llama-stack + image: llamastack/distribution-starter:latest + imagePullPolicy: Always # since we have specified latest instead of a version + env: + - name: ENABLE_CHROMADB + value: "true" + - name: CHROMADB_URL + value: http://chromadb.default.svc.cluster.local:6000 + - name: VLLM_URL + value: http://vllm-server.default.svc.cluster.local:8000/v1 + - name: VLLM_MAX_TOKENS + value: "3072" + - name: VLLM_SAFETY_URL + value: http://vllm-server-safety.default.svc.cluster.local:8001/v1 + - name: VLLM_TLS_VERIFY + value: "false" + - name: POSTGRES_HOST + value: postgres-server.default.svc.cluster.local + - name: POSTGRES_PORT + value: "5432" + - name: INFERENCE_MODEL + value: "${INFERENCE_MODEL}" + - name: SAFETY_MODEL + value: "${SAFETY_MODEL}" + - name: TAVILY_SEARCH_API_KEY + value: "${TAVILY_SEARCH_API_KEY}" + command: ["python", "-m", "llama_stack.core.server.server", "/etc/config/stack_run_config.yaml", "--port", "8321"] + ports: + - containerPort: 8321 + volumeMounts: + - name: llama-storage + mountPath: /root/.llama + - name: llama-config + mountPath: /etc/config + volumes: + - name: llama-storage + persistentVolumeClaim: + claimName: llama-pvc + - name: llama-config + configMap: + name: llama-stack-config diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/stack_run_config.yaml b/versioned_docs/version-v0.2.23/distributions/k8s/stack_run_config.yaml new file mode 100644 index 0000000..b841ab9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/stack_run_config.yaml @@ -0,0 +1,140 @@ +version: '2' +image_name: kubernetes-demo +apis: +- agents +- inference +- files +- safety +- telemetry +- tool_runtime +- vector_io +providers: + inference: + - provider_id: vllm-inference + provider_type: remote::vllm + config: + url: ${env.VLLM_URL:=http://localhost:8000/v1} + max_tokens: ${env.VLLM_MAX_TOKENS:=4096} + api_token: ${env.VLLM_API_TOKEN:=fake} + tls_verify: ${env.VLLM_TLS_VERIFY:=true} + - provider_id: vllm-safety + provider_type: remote::vllm + config: + url: ${env.VLLM_SAFETY_URL:=http://localhost:8000/v1} + max_tokens: ${env.VLLM_MAX_TOKENS:=4096} + api_token: ${env.VLLM_API_TOKEN:=fake} + tls_verify: ${env.VLLM_TLS_VERIFY:=true} + - provider_id: sentence-transformers + provider_type: inline::sentence-transformers + config: {} + vector_io: + - provider_id: ${env.ENABLE_CHROMADB:+chromadb} + provider_type: remote::chromadb + config: + url: ${env.CHROMADB_URL:=} + kvstore: + type: postgres + host: ${env.POSTGRES_HOST:=localhost} + port: ${env.POSTGRES_PORT:=5432} + db: ${env.POSTGRES_DB:=llamastack} + user: ${env.POSTGRES_USER:=llamastack} + password: ${env.POSTGRES_PASSWORD:=llamastack} + files: + - provider_id: meta-reference-files + provider_type: inline::localfs + config: + storage_dir: ${env.FILES_STORAGE_DIR:=~/.llama/distributions/starter/files} + metadata_store: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/distributions/starter}/files_metadata.db + safety: + - provider_id: llama-guard + provider_type: inline::llama-guard + config: + excluded_categories: [] + agents: + - provider_id: meta-reference + provider_type: inline::meta-reference + config: + persistence_store: + type: postgres + host: ${env.POSTGRES_HOST:=localhost} + port: ${env.POSTGRES_PORT:=5432} + db: ${env.POSTGRES_DB:=llamastack} + user: ${env.POSTGRES_USER:=llamastack} + password: ${env.POSTGRES_PASSWORD:=llamastack} + responses_store: + type: postgres + host: ${env.POSTGRES_HOST:=localhost} + port: ${env.POSTGRES_PORT:=5432} + db: ${env.POSTGRES_DB:=llamastack} + user: ${env.POSTGRES_USER:=llamastack} + password: ${env.POSTGRES_PASSWORD:=llamastack} + telemetry: + - provider_id: meta-reference + provider_type: inline::meta-reference + config: + service_name: "${env.OTEL_SERVICE_NAME:=\u200B}" + sinks: ${env.TELEMETRY_SINKS:=console} + tool_runtime: + - provider_id: brave-search + provider_type: remote::brave-search + config: + api_key: ${env.BRAVE_SEARCH_API_KEY:+} + max_results: 3 + - provider_id: tavily-search + provider_type: remote::tavily-search + config: + api_key: ${env.TAVILY_SEARCH_API_KEY:+} + max_results: 3 + - provider_id: rag-runtime + provider_type: inline::rag-runtime + config: {} + - provider_id: model-context-protocol + provider_type: remote::model-context-protocol + config: {} +metadata_store: + type: postgres + host: ${env.POSTGRES_HOST:=localhost} + port: ${env.POSTGRES_PORT:=5432} + db: ${env.POSTGRES_DB:=llamastack} + user: ${env.POSTGRES_USER:=llamastack} + password: ${env.POSTGRES_PASSWORD:=llamastack} + table_name: llamastack_kvstore +inference_store: + type: postgres + host: ${env.POSTGRES_HOST:=localhost} + port: ${env.POSTGRES_PORT:=5432} + db: ${env.POSTGRES_DB:=llamastack} + user: ${env.POSTGRES_USER:=llamastack} + password: ${env.POSTGRES_PASSWORD:=llamastack} +models: +- metadata: + embedding_dimension: 384 + model_id: all-MiniLM-L6-v2 + provider_id: sentence-transformers + model_type: embedding +- metadata: {} + model_id: ${env.INFERENCE_MODEL} + provider_id: vllm-inference + model_type: llm +- metadata: {} + model_id: ${env.SAFETY_MODEL:=meta-llama/Llama-Guard-3-1B} + provider_id: vllm-safety + model_type: llm +shields: +- shield_id: ${env.SAFETY_MODEL:=meta-llama/Llama-Guard-3-1B} +vector_dbs: [] +datasets: [] +scoring_fns: [] +benchmarks: [] +tool_groups: +- toolgroup_id: builtin::websearch + provider_id: tavily-search +- toolgroup_id: builtin::rag + provider_id: rag-runtime +server: + port: 8321 + auth: + provider_config: + type: github_token diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/ui-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/ui-k8s.yaml.template new file mode 100644 index 0000000..a6859cb --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/ui-k8s.yaml.template @@ -0,0 +1,68 @@ +apiVersion: apps/v1 +kind: Deployment +metadata: + name: llama-stack-ui + labels: + app.kubernetes.io/name: llama-stack + app.kubernetes.io/component: ui +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: llama-stack + app.kubernetes.io/component: ui + template: + metadata: + labels: + app.kubernetes.io/name: llama-stack + app.kubernetes.io/component: ui + spec: + containers: + - name: llama-stack-ui + image: node:18-alpine + command: ["/bin/sh"] + env: + - name: LLAMA_STACK_BACKEND_URL + value: "http://llama-stack-service:8321" + - name: LLAMA_STACK_UI_PORT + value: "8322" + - name: GITHUB_CLIENT_ID + value: "${GITHUB_CLIENT_ID}" + - name: GITHUB_CLIENT_SECRET + value: "${GITHUB_CLIENT_SECRET}" + - name: NEXTAUTH_URL + value: "${LLAMA_STACK_UI_URL}:8322" + args: + - -c + - | + # Install git (not included in alpine by default) + apk add --no-cache git + + # Clone the repository + echo "Cloning repository..." + git clone https://github.com/meta-llama/llama-stack.git /app + + # Navigate to the UI directory + echo "Navigating to UI directory..." + cd /app/llama_stack/ui + + # Check if package.json exists + if [ ! -f "package.json" ]; then + echo "ERROR: package.json not found in $(pwd)" + ls -la + exit 1 + fi + + # Install dependencies with verbose output + echo "Installing dependencies..." + npm install --verbose + + # Verify next is installed + echo "Checking if next is installed..." + npx next --version || echo "Next.js not found, checking node_modules..." + ls -la node_modules/.bin/ | grep next || echo "No next binary found" + + npm run dev + ports: + - containerPort: 8322 + workingDir: /app diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/vllm-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/vllm-k8s.yaml.template new file mode 100644 index 0000000..22bee4b --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/vllm-k8s.yaml.template @@ -0,0 +1,70 @@ +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: vllm-models +spec: + accessModes: + - ReadWriteOnce + volumeMode: Filesystem + resources: + requests: + storage: 50Gi +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: vllm-server +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: vllm + template: + metadata: + labels: + app.kubernetes.io/name: vllm + workload-type: inference + spec: + nodeSelector: + eks.amazonaws.com/nodegroup: gpu + containers: + - name: vllm + image: vllm/vllm-openai:latest + command: ["/bin/sh", "-c"] + args: + - "vllm serve ${INFERENCE_MODEL} --dtype float16 --enforce-eager --max-model-len 4096 --gpu-memory-utilization 0.6 --enable-auto-tool-choice --tool-call-parser llama4_pythonic" + env: + - name: INFERENCE_MODEL + value: "${INFERENCE_MODEL}" + - name: HUGGING_FACE_HUB_TOKEN + valueFrom: + secretKeyRef: + name: hf-token-secret + key: token + ports: + - containerPort: 8000 + resources: + limits: + nvidia.com/gpu: 1 + requests: + nvidia.com/gpu: 1 + volumeMounts: + - name: llama-storage + mountPath: /root/.cache/huggingface + volumes: + - name: llama-storage + persistentVolumeClaim: + claimName: vllm-models +--- +apiVersion: v1 +kind: Service +metadata: + name: vllm-server +spec: + selector: + app.kubernetes.io/name: vllm + ports: + - protocol: TCP + port: 8000 + targetPort: 8000 + type: ClusterIP diff --git a/versioned_docs/version-v0.2.23/distributions/k8s/vllm-safety-k8s.yaml.template b/versioned_docs/version-v0.2.23/distributions/k8s/vllm-safety-k8s.yaml.template new file mode 100644 index 0000000..37b2b9a --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/k8s/vllm-safety-k8s.yaml.template @@ -0,0 +1,71 @@ +apiVersion: v1 +kind: PersistentVolumeClaim +metadata: + name: vllm-models-safety +spec: + accessModes: + - ReadWriteOnce + volumeMode: Filesystem + resources: + requests: + storage: 30Gi +--- +apiVersion: apps/v1 +kind: Deployment +metadata: + name: vllm-server-safety +spec: + replicas: 1 + selector: + matchLabels: + app.kubernetes.io/name: vllm-safety + template: + metadata: + labels: + app.kubernetes.io/name: vllm-safety + workload-type: inference + spec: + nodeSelector: + eks.amazonaws.com/nodegroup: gpu + containers: + - name: vllm-safety + image: vllm/vllm-openai:latest + command: ["/bin/sh", "-c"] + args: [ + "vllm serve ${SAFETY_MODEL} --dtype float16 --enforce-eager --max-model-len 4096 --port 8001 --gpu-memory-utilization 0.3" + ] + env: + - name: SAFETY_MODEL + value: "${SAFETY_MODEL}" + - name: HUGGING_FACE_HUB_TOKEN + valueFrom: + secretKeyRef: + name: hf-token-secret + key: token + ports: + - containerPort: 8001 + resources: + limits: + nvidia.com/gpu: 1 + requests: + nvidia.com/gpu: 1 + volumeMounts: + - name: llama-storage + mountPath: /root/.cache/huggingface + volumes: + - name: llama-storage + persistentVolumeClaim: + claimName: vllm-models-safety +--- +apiVersion: v1 +kind: Service +metadata: + name: vllm-server-safety +spec: + selector: + app.kubernetes.io/name: vllm-safety + ports: + - protocol: TCP + port: 8001 + targetPort: 8001 + type: ClusterIP diff --git a/versioned_docs/version-v0.2.23/distributions/list_of_distributions.mdx b/versioned_docs/version-v0.2.23/distributions/list_of_distributions.mdx new file mode 100644 index 0000000..813d3c7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/list_of_distributions.mdx @@ -0,0 +1,134 @@ +--- +title: Available Distributions +description: List of available distributions for Llama Stack +sidebar_label: Available Distributions +sidebar_position: 2 +--- + +# Available Distributions + +Llama Stack provides several pre-configured distributions to help you get started quickly. Choose the distribution that best fits your hardware and use case. + +## Quick Reference + +| Distribution | Use Case | Hardware Requirements | Provider | +|--------------|----------|----------------------|----------| +| `distribution-starter` | General purpose, prototyping | Any (CPU/GPU) | Ollama, Remote APIs | +| `distribution-meta-reference-gpu` | High-performance inference | GPU required | Local GPU inference | +| Remote-hosted | Production, managed service | None | Partner providers | +| iOS/Android SDK | Mobile applications | Mobile device | On-device inference | + +## Choose Your Distribution + +### ๐Ÿš€ Getting Started (Recommended for Beginners) + +**Use `distribution-starter` if you want to:** +- Prototype quickly without GPU requirements +- Use remote inference providers (Fireworks, Together, vLLM etc.) +- Run locally with Ollama for development + +```bash +docker pull llama-stack/distribution-starter +``` + +**Guides:** [Starter Distribution Guide](self_hosted_distro/starter) + +### ๐Ÿ–ฅ๏ธ Self-Hosted with GPU + +**Use `distribution-meta-reference-gpu` if you:** +- Have access to GPU hardware +- Want maximum performance and control +- Need to run inference locally + +```bash +docker pull llama-stack/distribution-meta-reference-gpu +``` + +**Guides:** [Meta Reference GPU Guide](self_hosted_distro/meta-reference-gpu) + +### ๐Ÿ–ฅ๏ธ Self-Hosted with NVIDA NeMo Microservices + +**Use `nvidia` if you:** +- Want to use Llama Stack with NVIDIA NeMo Microservices + +**Guides:** [NVIDIA Distribution Guide](self_hosted_distro/nvidia) + +### โ˜๏ธ Managed Hosting + +**Use remote-hosted endpoints if you:** +- Don't want to manage infrastructure +- Need production-ready reliability +- Prefer managed services + +**Partners:** [Fireworks.ai](https://fireworks.ai) and [Together.xyz](https://together.xyz) + +**Guides:** [Remote-Hosted Endpoints](./remote_hosted_distro/) + +### ๐Ÿ“ฑ Mobile Development + +**Use mobile SDKs if you:** +- Are building iOS or Android applications +- Need on-device inference capabilities +- Want offline functionality + +- [iOS SDK](ondevice_distro/ios_sdk) +- [Android SDK](ondevice_distro/android_sdk) + +### ๐Ÿ”ง Custom Solutions + +**Build your own distribution if:** +- None of the above fit your specific needs +- You need custom configurations +- You want to optimize for your specific use case + +**Guides:** [Building Custom Distributions](./building_distro) + +## Detailed Documentation + +### Self-Hosted Distributions + +```{toctree} +:maxdepth: 1 + +self_hosted_distro/starter +self_hosted_distro/meta-reference-gpu +``` + +### Remote-Hosted Solutions + +```{toctree} +:maxdepth: 1 + +remote_hosted_distro/index +``` + +### Mobile SDKs + +```{toctree} +:maxdepth: 1 + +ondevice_distro/ios_sdk +ondevice_distro/android_sdk +``` + +## Decision Flow + +```mermaid +graph TD + A[What's your use case?] --> B{Need mobile app?} + B -->|Yes| C[Use Mobile SDKs] + B -->|No| D{Have GPU hardware?} + D -->|Yes| E[Use Meta Reference GPU] + D -->|No| F{Want managed hosting?} + F -->|Yes| G[Use Remote-Hosted] + F -->|No| H[Use Starter Distribution] +``` + +## Next Steps + +1. **Choose your distribution** from the options above +2. **Follow the setup guide** for your selected distribution +3. **Configure your providers** with API keys or local models +4. **Start building** with Llama Stack! + +For help choosing or troubleshooting, check our [Getting Started Guide](/docs/getting_started/quickstart) or [Community Support](https://github.com/llama-stack/llama-stack/discussions). diff --git a/versioned_docs/version-v0.2.23/distributions/ondevice_distro/android_sdk.md b/versioned_docs/version-v0.2.23/distributions/ondevice_distro/android_sdk.md new file mode 100644 index 0000000..bfa294e --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/ondevice_distro/android_sdk.md @@ -0,0 +1,262 @@ +# Llama Stack Client Kotlin API Library + +We are excited to share a guide for a Kotlin Library that brings front the benefits of Llama Stack to your Android device. This library is a set of SDKs that provide a simple and effective way to integrate AI capabilities into your Android app whether it is local (on-device) or remote inference. + +Features: +- Local Inferencing: Run Llama models purely on-device with real-time processing. We currently utilize ExecuTorch as the local inference distributor and may support others in the future. + - [ExecuTorch](https://github.com/pytorch/executorch/tree/main) is a complete end-to-end solution within the PyTorch framework for inferencing capabilities on-device with high portability and seamless performance. +- Remote Inferencing: Perform inferencing tasks remotely with Llama models hosted on a remote connection (or serverless localhost). +- Simple Integration: With easy-to-use APIs, a developer can quickly integrate Llama Stack in their Android app. The difference with local vs remote inferencing is also minimal. + +Latest Release Notes: [link](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release) + +*Tagged releases are stable versions of the project. While we strive to maintain a stable main branch, it's not guaranteed to be free of bugs or issues.* + +## Android Demo App +Check out our demo app to see how to integrate Llama Stack into your Android app: [Android Demo App](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release/examples/android_app) + +The key files in the app are `ExampleLlamaStackLocalInference.kt`, `ExampleLlamaStackRemoteInference.kts`, and `MainActivity.java`. With encompassed business logic, the app shows how to use Llama Stack for both the environments. + +## Quick Start + +### Add Dependencies +#### Kotlin Library +Add the following dependency in your `build.gradle.kts` file: +``` +dependencies { + implementation("com.llama.llamastack:llama-stack-client-kotlin:0.2.2") +} +``` +This will download jar files in your gradle cache in a directory like `~/.gradle/caches/modules-2/files-2.1/com.llama.llamastack/` + +If you plan on doing remote inferencing this is sufficient to get started. + +#### Dependency for Local + +For local inferencing, it is required to include the ExecuTorch library into your app. + +Include the ExecuTorch library by: +1. Download the `download-prebuilt-et-lib.sh` script file from the [llama-stack-client-kotlin-client-local](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release/llama-stack-client-kotlin-client-local/download-prebuilt-et-lib.sh) directory to your local machine. +2. Move the script to the top level of your Android app where the `app` directory resides. +3. Run `sh download-prebuilt-et-lib.sh` to create an `app/libs` directory and download the `executorch.aar` in that path. This generates an ExecuTorch library for the XNNPACK delegate. +4. Add the `executorch.aar` dependency in your `build.gradle.kts` file: +``` +dependencies { + ... + implementation(files("libs/executorch.aar")) + ... +} +``` + +See other dependencies for the local RAG in Android app [README](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release/examples/android_app#quick-start). + +## Llama Stack APIs in Your Android App +Breaking down the demo app, this section will show the core pieces that are used to initialize and run inference with Llama Stack using the Kotlin library. + +### Setup Remote Inferencing +Start a Llama Stack server on localhost. Here is an example of how you can do this using the firework.ai distribution: +``` +uv venv starter --python 3.12 +source starter/bin/activate # On Windows: starter\Scripts\activate +pip install --no-cache llama-stack==0.2.2 +llama stack build --distro starter --image-type venv +export FIREWORKS_API_KEY= +llama stack run starter --port 5050 +``` + +Ensure the Llama Stack server version is the same as the Kotlin SDK Library for maximum compatibility. + +Other inference providers: [Table](/docs/) + +How to set remote localhost in Demo App: [Settings](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release/examples/android_app#settings) + +### Initialize the Client +A client serves as the primary interface for interacting with a specific inference type and its associated parameters. Only after client is initialized then you can configure and start inferences. + + + + + + + + + + +
Local InferenceRemote Inference
+ +``` +client = LlamaStackClientLocalClient + .builder() + .modelPath(modelPath) + .tokenizerPath(tokenizerPath) + .temperature(temperature) + .build() +``` + + +``` +// remoteURL is a string like "http://localhost:5050" +client = LlamaStackClientOkHttpClient + .builder() + .baseUrl(remoteURL) + .build() +``` +
+ + +### Run Inference +With the Kotlin Library managing all the major operational logic, there are minimal to no changes when running simple chat inference for local or remote: + +``` +val result = client!!.inference().chatCompletion( + InferenceChatCompletionParams.builder() + .modelId(modelName) + .messages(listOfMessages) + .build() + ) + +// response contains string with response from model +var response = result.asChatCompletionResponse().completionMessage().content().string(); +``` + +[Remote only] For inference with a streaming response: + +``` +val result = client!!.inference().chatCompletionStreaming( + InferenceChatCompletionParams.builder() + .modelId(modelName) + .messages(listOfMessages) + .build() + ) + +// Response can be received as a asChatCompletionResponseStreamChunk as part of a callback. +// See Android demo app for a detailed implementation example. +``` + +### Setup Custom Tool Calling + +Android demo app for more details: [Custom Tool Calling](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release/examples/android_app#tool-calling) + +## Advanced Users + +The purpose of this section is to share more details with users that would like to dive deeper into the Llama Stack Kotlin Library. Whether youโ€™re interested in contributing to the open source library, debugging or just want to learn more, this section is for you! + +### Prerequisite + +You must complete the following steps: +1. Clone the repo (`git clone https://github.com/meta-llama/llama-stack-client-kotlin.git -b latest-release`) +2. Port the appropriate ExecuTorch libraries over into your Llama Stack Kotlin library environment. +``` +cd llama-stack-client-kotlin-client-local +sh download-prebuilt-et-lib.sh --unzip +``` + +Now you will notice that the `jni/` , `libs/`, and `AndroidManifest.xml` files from the `executorch.aar` file are present in the local module. This way the local client module will be able to realize the ExecuTorch SDK. + +### Building for Development/Debugging +If youโ€™d like to contribute to the Kotlin library via development, debug, or add play around with the library with various print statements, run the following command in your terminal under the llama-stack-client-kotlin directory. + +``` +sh build-libs.sh +``` + +Output: .jar files located in the build-jars directory + +Copy the .jar files over to the lib directory in your Android app. At the same time make sure to remove the llama-stack-client-kotlin dependency within your build.gradle.kts file in your app (or if you are using the demo app) to avoid having multiple llama stack client dependencies. + +### Additional Options for Local Inferencing +Currently we provide additional properties support with local inferencing. In order to get the tokens/sec metric for each inference call, add the following code in your Android app after you run your chatCompletion inference function. The Reference app has this implementation as well: +``` +var tps = (result.asChatCompletionResponse()._additionalProperties()["tps"] as JsonNumber).value as Float +``` +We will be adding more properties in the future. + +### Additional Options for Remote Inferencing + +#### Network options + +##### Retries + +Requests that experience certain errors are automatically retried 2 times by default, with a short exponential backoff. Connection errors (for example, due to a network connectivity problem), 408 Request Timeout, 409 Conflict, 429 Rate Limit, and >=500 Internal errors will all be retried by default. +You can provide a `maxRetries` on the client builder to configure this: + +```kotlin +val client = LlamaStackClientOkHttpClient.builder() + .fromEnv() + .maxRetries(4) + .build() +``` + +##### Timeouts + +Requests time out after 1 minute by default. You can configure this on the client builder: + +```kotlin +val client = LlamaStackClientOkHttpClient.builder() + .fromEnv() + .timeout(Duration.ofSeconds(30)) + .build() +``` + +##### Proxies + +Requests can be routed through a proxy. You can configure this on the client builder: + +```kotlin +val client = LlamaStackClientOkHttpClient.builder() + .fromEnv() + .proxy(new Proxy( + Type.HTTP, + new InetSocketAddress("proxy.com", 8080) + )) + .build() +``` + +##### Environments + +Requests are made to the production environment by default. You can connect to other environments, like `sandbox`, via the client builder: + +```kotlin +val client = LlamaStackClientOkHttpClient.builder() + .fromEnv() + .sandbox() + .build() +``` + +### Error Handling +This library throws exceptions in a single hierarchy for easy handling: + +- **`LlamaStackClientException`** - Base exception for all exceptions + + - **`LlamaStackClientServiceException`** - HTTP errors with a well-formed response body we were able to parse. The exception message and the `.debuggingRequestId()` will be set by the server. + + | 400 | BadRequestException | + | ------ | ----------------------------- | + | 401 | AuthenticationException | + | 403 | PermissionDeniedException | + | 404 | NotFoundException | + | 422 | UnprocessableEntityException | + | 429 | RateLimitException | + | 5xx | InternalServerException | + | others | UnexpectedStatusCodeException | + + - **`LlamaStackClientIoException`** - I/O networking errors + - **`LlamaStackClientInvalidDataException`** - any other exceptions on the client side, e.g.: + - We failed to serialize the request body + - We failed to parse the response body (has access to response code and body) + +## Reporting Issues +If you encountered any bugs or issues following this guide please file a bug/issue on our [Github issue tracker](https://github.com/meta-llama/llama-stack-client-kotlin/issues). + +## Known Issues +We're aware of the following issues and are working to resolve them: +1. Streaming response is a work-in-progress for local and remote inference +2. Due to #1, agents are not supported at the time. LS agents only work in streaming mode +3. Changing to another model is a work in progress for local and remote platforms + +## Thanks +We'd like to extend our thanks to the ExecuTorch team for providing their support as we integrated ExecuTorch as one of the local inference distributors for Llama Stack. Checkout [ExecuTorch Github repo](https://github.com/pytorch/executorch/tree/main) for more information. + +--- + +The API interface is generated using the OpenAPI standard with [Stainless](https://www.stainlessapi.com/). diff --git a/versioned_docs/version-v0.2.23/distributions/ondevice_distro/ios_sdk.md b/versioned_docs/version-v0.2.23/distributions/ondevice_distro/ios_sdk.md new file mode 100644 index 0000000..de4002e --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/ondevice_distro/ios_sdk.md @@ -0,0 +1,134 @@ +# iOS SDK + +We offer both remote and on-device use of Llama Stack in Swift via a single SDK [llama-stack-client-swift](https://github.com/meta-llama/llama-stack-client-swift/) that contains two components: +1. LlamaStackClient for remote +2. Local Inference for on-device + +```{image} ../../../_static/remote_or_local.gif +:alt: Seamlessly switching between local, on-device inference and remote hosted inference +:width: 412px +:align: center +``` + +## Remote Only + +If you don't want to run inference on-device, then you can connect to any hosted Llama Stack distribution with #1. + +1. Add `https://github.com/meta-llama/llama-stack-client-swift/` as a Package Dependency in Xcode + +2. Add `LlamaStackClient` as a framework to your app target + +3. Call an API: + +```swift +import LlamaStackClient + +let agents = RemoteAgents(url: URL(string: "http://localhost:8321")!) +let request = Components.Schemas.CreateAgentTurnRequest( + agent_id: agentId, + messages: [ + .UserMessage(Components.Schemas.UserMessage( + content: .case1("Hello Llama!"), + role: .user + )) + ], + session_id: self.agenticSystemSessionId, + stream: true + ) + + for try await chunk in try await agents.createTurn(request: request) { + let payload = chunk.event.payload + // ... +``` + +Check out [iOSCalendarAssistant](https://github.com/meta-llama/llama-stack-client-swift/tree/main/examples/ios_calendar_assistant) for a complete app demo. + +## LocalInference + +LocalInference provides a local inference implementation powered by [executorch](https://github.com/pytorch/executorch/). + +Llama Stack currently supports on-device inference for iOS with Android coming soon. You can run on-device inference on Android today using [executorch](https://github.com/pytorch/executorch/tree/main/examples/demo-apps/android/LlamaDemo), PyTorchโ€™s on-device inference library. + +The APIs *work the same as remote* โ€“ย the only difference is you'll instead use the `LocalAgents` / `LocalInference` classes and pass in a `DispatchQueue`: + +```swift +private let runnerQueue = DispatchQueue(label: "org.llamastack.stacksummary") +let inference = LocalInference(queue: runnerQueue) +let agents = LocalAgents(inference: self.inference) +``` + +Check out [iOSCalendarAssistantWithLocalInf](https://github.com/meta-llama/llama-stack-client-swift/tree/main/examples/ios_calendar_assistant) for a complete app demo. + +### Installation + +We're working on making LocalInference easier to set up.ย For now, you'll need to import it via `.xcframework`: + +1. Clone the executorch submodule in this repo and its dependencies: `git submodule update --init --recursive` +1. Install [Cmake](https://cmake.org/) for the executorch build` +1. Drag `LocalInference.xcodeproj` into your project +1. Add `LocalInference` as a framework in your app target + +### Preparing a model + +1. Prepare a `.pte` file [following the executorch docs](https://github.com/pytorch/executorch/blob/main/examples/models/llama/README.md#step-2-prepare-model) +2. Bundle the `.pte` and `tokenizer.model` file into your app + +We now support models quantized using SpinQuant and QAT-LoRA which offer a significant performance boost (demo app on iPhone 13 Pro): + + +| Llama 3.2 1B | Tokens / Second (total) | | Time-to-First-Token (sec) | | +| :---- | :---- | :---- | :---- | :---- | +| | Haiku | Paragraph | Haiku | Paragraph | +| BF16 | 2.2 | 2.5 | 2.3 | 1.9 | +| QAT+LoRA | 7.1 | 3.3 | 0.37 | 0.24 | +| SpinQuant | 10.1 | 5.2 | 0.2 | 0.2 | + + +### Using LocalInference + +1. Instantiate LocalInference with a DispatchQueue. Optionally, pass it into your agents service: + +```swift + init () { + runnerQueue = DispatchQueue(label: "org.meta.llamastack") + inferenceService = LocalInferenceService(queue: runnerQueue) + agentsService = LocalAgentsService(inference: inferenceService) + } +``` + +2. Before making any inference calls, load your model from your bundle: + +```swift +let mainBundle = Bundle.main +inferenceService.loadModel( + modelPath: mainBundle.url(forResource: "llama32_1b_spinquant", withExtension: "pte"), + tokenizerPath: mainBundle.url(forResource: "tokenizer", withExtension: "model"), + completion: {_ in } // use to handle load failures +) +``` + +3. Make inference calls (or agents calls) as you normally would with LlamaStack: + +``` +for await chunk in try await agentsService.initAndCreateTurn( + messages: [ + .UserMessage(Components.Schemas.UserMessage( + content: .case1("Call functions as needed to handle any actions in the following text:\n\n" + text), + role: .user)) + ] +) { +``` + +### Troubleshooting + +If you receive errors like "missing package product" or "invalid checksum", try cleaning the build folder and resetting the Swift package cache: + +(Opt+Click) Product > Clean Build Folder Immediately + +``` +rm -rf \ + ~/Library/org.swift.swiftpm \ + ~/Library/Caches/org.swift.swiftpm \ + ~/Library/Caches/com.apple.dt.Xcode \ + ~/Library/Developer/Xcode/DerivedData +``` diff --git a/versioned_docs/version-v0.2.23/distributions/remote_hosted_distro/index.mdx b/versioned_docs/version-v0.2.23/distributions/remote_hosted_distro/index.mdx new file mode 100644 index 0000000..ef5a83d --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/remote_hosted_distro/index.mdx @@ -0,0 +1,20 @@ +# Remote-Hosted Distributions + +Remote-Hosted distributions are available endpoints serving Llama Stack API that you can directly connect to. + +| Distribution | Endpoint | Inference | Agents | Memory | Safety | Telemetry | +|-------------|----------|-----------|---------|---------|---------|------------| +| Together | [https://llama-stack.together.ai](https://llama-stack.together.ai) | remote::together | meta-reference | remote::weaviate | meta-reference | meta-reference | +| Fireworks | [https://llamastack-preview.fireworks.ai](https://llamastack-preview.fireworks.ai) | remote::fireworks | meta-reference | remote::weaviate | meta-reference | meta-reference | + +## Connecting to Remote-Hosted Distributions + +You can use `llama-stack-client` to interact with these endpoints. For example, to list the available models served by the Fireworks endpoint: + +```bash +$ pip install llama-stack-client +$ llama-stack-client configure --endpoint https://llamastack-preview.fireworks.ai +$ llama-stack-client models list +``` + +Checkout the [llama-stack-client-python](https://github.com/meta-llama/llama-stack-client-python/blob/main/docs/cli_reference.md) repo for more details on how to use the `llama-stack-client` CLI. Checkout [llama-stack-app](https://github.com/meta-llama/llama-stack-apps/tree/main) for examples applications built on top of Llama Stack. diff --git a/versioned_docs/version-v0.2.23/distributions/remote_hosted_distro/watsonx.md b/versioned_docs/version-v0.2.23/distributions/remote_hosted_distro/watsonx.md new file mode 100644 index 0000000..977af90 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/remote_hosted_distro/watsonx.md @@ -0,0 +1,78 @@ +--- +orphan: true +--- + +# watsonx Distribution + +```{toctree} +:maxdepth: 2 +:hidden: + +self +``` + +The `llamastack/distribution-watsonx` distribution consists of the following provider configurations. + +| API | Provider(s) | +|-----|-------------| +| agents | `inline::meta-reference` | +| datasetio | `remote::huggingface`, `inline::localfs` | +| eval | `inline::meta-reference` | +| inference | `remote::watsonx`, `inline::sentence-transformers` | +| safety | `inline::llama-guard` | +| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | +| telemetry | `inline::meta-reference` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::rag-runtime`, `remote::model-context-protocol` | +| vector_io | `inline::faiss` | + + + +### Environment Variables + +The following environment variables can be configured: + +- `LLAMASTACK_PORT`: Port for the Llama Stack distribution server (default: `5001`) +- `WATSONX_API_KEY`: watsonx API Key (default: ``) +- `WATSONX_PROJECT_ID`: watsonx Project ID (default: ``) + +### Models + +The following models are available by default: + +- `meta-llama/llama-3-3-70b-instruct (aliases: meta-llama/Llama-3.3-70B-Instruct)` +- `meta-llama/llama-2-13b-chat (aliases: meta-llama/Llama-2-13b)` +- `meta-llama/llama-3-1-70b-instruct (aliases: meta-llama/Llama-3.1-70B-Instruct)` +- `meta-llama/llama-3-1-8b-instruct (aliases: meta-llama/Llama-3.1-8B-Instruct)` +- `meta-llama/llama-3-2-11b-vision-instruct (aliases: meta-llama/Llama-3.2-11B-Vision-Instruct)` +- `meta-llama/llama-3-2-1b-instruct (aliases: meta-llama/Llama-3.2-1B-Instruct)` +- `meta-llama/llama-3-2-3b-instruct (aliases: meta-llama/Llama-3.2-3B-Instruct)` +- `meta-llama/llama-3-2-90b-vision-instruct (aliases: meta-llama/Llama-3.2-90B-Vision-Instruct)` +- `meta-llama/llama-guard-3-11b-vision (aliases: meta-llama/Llama-Guard-3-11B-Vision)` + + +### Prerequisite: API Keys + +Make sure you have access to a watsonx API Key. You can get one by referring [watsonx.ai](https://www.ibm.com/docs/en/masv-and-l/maximo-manage/continuous-delivery?topic=setup-create-watsonx-api-key). + + +## Running Llama Stack with watsonx + +You can do this via venv or Docker which has a pre-built image. + +### Via Docker + +This method allows you to get started quickly without having to build the distribution code. + +```bash +LLAMA_STACK_PORT=5001 +docker run \ + -it \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ./run.yaml:/root/my-run.yaml \ + llamastack/distribution-watsonx \ + --config /root/my-run.yaml \ + --port $LLAMA_STACK_PORT \ + --env WATSONX_API_KEY=$WATSONX_API_KEY \ + --env WATSONX_PROJECT_ID=$WATSONX_PROJECT_ID \ + --env WATSONX_BASE_URL=$WATSONX_BASE_URL +``` diff --git a/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/dell-tgi.md b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/dell-tgi.md new file mode 100644 index 0000000..5fca297 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/dell-tgi.md @@ -0,0 +1,78 @@ +--- +orphan: true +--- +# Dell-TGI Distribution + +```{toctree} +:maxdepth: 2 +:hidden: + +self +``` + +The `llamastack/distribution-tgi` distribution consists of the following provider configurations. + + +| **API** | **Inference** | **Agents** | **Memory** | **Safety** | **Telemetry** | +|----------------- |--------------- |---------------- |-------------------------------------------------- |---------------- |---------------- | +| **Provider(s)** | remote::tgi | meta-reference | meta-reference, remote::pgvector, remote::chroma | meta-reference | meta-reference | + + +The only difference vs. the `tgi` distribution is that it runs the Dell-TGI server for inference. + + +### Start the Distribution (Single Node GPU) + +> [!NOTE] +> This assumes you have access to GPU to start a TGI server with access to your GPU. + +``` +$ cd distributions/dell-tgi/ +$ ls +compose.yaml README.md run.yaml +$ docker compose up +``` + +The script will first start up TGI server, then start up Llama Stack distribution server hooking up to the remote TGI provider for inference. You should be able to see the following outputs -- +``` +[text-generation-inference] | 2024-10-15T18:56:33.810397Z INFO text_generation_router::server: router/src/server.rs:1813: Using config Some(Llama) +[text-generation-inference] | 2024-10-15T18:56:33.810448Z WARN text_generation_router::server: router/src/server.rs:1960: Invalid hostname, defaulting to 0.0.0.0 +[text-generation-inference] | 2024-10-15T18:56:33.864143Z INFO text_generation_router::server: router/src/server.rs:2353: Connected +INFO: Started server process [1] +INFO: Waiting for application startup. +INFO: Application startup complete. +INFO: Uvicorn running on http://[::]:8321 (Press CTRL+C to quit) +``` + +To kill the server +``` +docker compose down +``` + +### (Alternative) Dell-TGI server + llama stack run (Single Node GPU) + +#### Start Dell-TGI server locally +``` +docker run -it --pull always --shm-size 1g -p 80:80 --gpus 4 \ +-e NUM_SHARD=4 +-e MAX_BATCH_PREFILL_TOKENS=32768 \ +-e MAX_INPUT_TOKENS=8000 \ +-e MAX_TOTAL_TOKENS=8192 \ +registry.dell.huggingface.co/enterprise-dell-inference-meta-llama-meta-llama-3.1-8b-instruct +``` + + +#### Start Llama Stack server pointing to TGI server + +``` +docker run --pull always --network host -it -p 8321:8321 -v ./run.yaml:/root/my-run.yaml --gpus=all llamastack/distribution-tgi --yaml_config /root/my-run.yaml +``` + +Make sure in you `run.yaml` file, you inference provider is pointing to the correct TGI server endpoint. E.g. +``` +inference: + - provider_id: tgi0 + provider_type: remote::tgi + config: + url: http://127.0.0.1:5009 +``` diff --git a/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/dell.md b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/dell.md new file mode 100644 index 0000000..68e7b6f --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/dell.md @@ -0,0 +1,190 @@ +--- +orphan: true +--- + + +# Dell Distribution of Llama Stack + +```{toctree} +:maxdepth: 2 +:hidden: + +self +``` + +The `llamastack/distribution-dell` distribution consists of the following provider configurations. + +| API | Provider(s) | +|-----|-------------| +| agents | `inline::meta-reference` | +| datasetio | `remote::huggingface`, `inline::localfs` | +| eval | `inline::meta-reference` | +| inference | `remote::tgi`, `inline::sentence-transformers` | +| safety | `inline::llama-guard` | +| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | +| telemetry | `inline::meta-reference` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::rag-runtime` | +| vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | + + +You can use this distribution if you have GPUs and want to run an independent TGI or Dell Enterprise Hub container for running inference. + +### Environment Variables + +The following environment variables can be configured: + +- `DEH_URL`: URL for the Dell inference server (default: `http://0.0.0.0:8181`) +- `DEH_SAFETY_URL`: URL for the Dell safety inference server (default: `http://0.0.0.0:8282`) +- `CHROMA_URL`: URL for the Chroma server (default: `http://localhost:6601`) +- `INFERENCE_MODEL`: Inference model loaded into the TGI server (default: `meta-llama/Llama-3.2-3B-Instruct`) +- `SAFETY_MODEL`: Name of the safety (Llama-Guard) model to use (default: `meta-llama/Llama-Guard-3-1B`) + + +## Setting up Inference server using Dell Enterprise Hub's custom TGI container. + +NOTE: This is a placeholder to run inference with TGI. This will be updated to use [Dell Enterprise Hub's containers](https://dell.huggingface.co/authenticated/models) once verified. + +```bash +export INFERENCE_PORT=8181 +export DEH_URL=http://0.0.0.0:$INFERENCE_PORT +export INFERENCE_MODEL=meta-llama/Llama-3.1-8B-Instruct +export CHROMADB_HOST=localhost +export CHROMADB_PORT=6601 +export CHROMA_URL=http://$CHROMADB_HOST:$CHROMADB_PORT +export CUDA_VISIBLE_DEVICES=0 +export LLAMA_STACK_PORT=8321 + +docker run --rm -it \ + --pull always \ + --network host \ + -v $HOME/.cache/huggingface:/data \ + -e HF_TOKEN=$HF_TOKEN \ + -p $INFERENCE_PORT:$INFERENCE_PORT \ + --gpus $CUDA_VISIBLE_DEVICES \ + ghcr.io/huggingface/text-generation-inference \ + --dtype bfloat16 \ + --usage-stats off \ + --sharded false \ + --cuda-memory-fraction 0.7 \ + --model-id $INFERENCE_MODEL \ + --port $INFERENCE_PORT --hostname 0.0.0.0 +``` + +If you are using Llama Stack Safety / Shield APIs, then you will need to also run another instance of a TGI with a corresponding safety model like `meta-llama/Llama-Guard-3-1B` using a script like: + +```bash +export SAFETY_INFERENCE_PORT=8282 +export DEH_SAFETY_URL=http://0.0.0.0:$SAFETY_INFERENCE_PORT +export SAFETY_MODEL=meta-llama/Llama-Guard-3-1B +export CUDA_VISIBLE_DEVICES=1 + +docker run --rm -it \ + --pull always \ + --network host \ + -v $HOME/.cache/huggingface:/data \ + -e HF_TOKEN=$HF_TOKEN \ + -p $SAFETY_INFERENCE_PORT:$SAFETY_INFERENCE_PORT \ + --gpus $CUDA_VISIBLE_DEVICES \ + ghcr.io/huggingface/text-generation-inference \ + --dtype bfloat16 \ + --usage-stats off \ + --sharded false \ + --cuda-memory-fraction 0.7 \ + --model-id $SAFETY_MODEL \ + --hostname 0.0.0.0 \ + --port $SAFETY_INFERENCE_PORT +``` + +## Dell distribution relies on ChromaDB for vector database usage + +You can start a chroma-db easily using docker. +```bash +# This is where the indices are persisted +mkdir -p $HOME/chromadb + +podman run --rm -it \ + --network host \ + --name chromadb \ + -v $HOME/chromadb:/chroma/chroma \ + -e IS_PERSISTENT=TRUE \ + chromadb/chroma:latest \ + --port $CHROMADB_PORT \ + --host $CHROMADB_HOST +``` + +## Running Llama Stack + +Now you are ready to run Llama Stack with TGI as the inference provider. You can do this via venv or Docker which has a pre-built image. + +### Via Docker + +This method allows you to get started quickly without having to build the distribution code. + +```bash +docker run -it \ + --pull always \ + --network host \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v $HOME/.llama:/root/.llama \ + # NOTE: mount the llama-stack / llama-model directories if testing local changes else not needed + -v /home/hjshah/git/llama-stack:/app/llama-stack-source -v /home/hjshah/git/llama-models:/app/llama-models-source \ + # localhost/distribution-dell:dev if building / testing locally + llamastack/distribution-dell\ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=$INFERENCE_MODEL \ + --env DEH_URL=$DEH_URL \ + --env CHROMA_URL=$CHROMA_URL + +``` + +If you are using Llama Stack Safety / Shield APIs, use: + +```bash +# You need a local checkout of llama-stack to run this, get it using +# git clone https://github.com/meta-llama/llama-stack.git +cd /path/to/llama-stack + +export SAFETY_INFERENCE_PORT=8282 +export DEH_SAFETY_URL=http://0.0.0.0:$SAFETY_INFERENCE_PORT +export SAFETY_MODEL=meta-llama/Llama-Guard-3-1B + +docker run \ + -it \ + --pull always \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v $HOME/.llama:/root/.llama \ + -v ./llama_stack/distributions/tgi/run-with-safety.yaml:/root/my-run.yaml \ + llamastack/distribution-dell \ + --config /root/my-run.yaml \ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=$INFERENCE_MODEL \ + --env DEH_URL=$DEH_URL \ + --env SAFETY_MODEL=$SAFETY_MODEL \ + --env DEH_SAFETY_URL=$DEH_SAFETY_URL \ + --env CHROMA_URL=$CHROMA_URL +``` + +### Via venv + +Make sure you have done `pip install llama-stack` and have the Llama Stack CLI available. + +```bash +llama stack build --distro dell --image-type venv +llama stack run dell + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=$INFERENCE_MODEL \ + --env DEH_URL=$DEH_URL \ + --env CHROMA_URL=$CHROMA_URL +``` + +If you are using Llama Stack Safety / Shield APIs, use: + +```bash +llama stack run ./run-with-safety.yaml \ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=$INFERENCE_MODEL \ + --env DEH_URL=$DEH_URL \ + --env SAFETY_MODEL=$SAFETY_MODEL \ + --env DEH_SAFETY_URL=$DEH_SAFETY_URL \ + --env CHROMA_URL=$CHROMA_URL +``` diff --git a/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/meta-reference-gpu.md b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/meta-reference-gpu.md new file mode 100644 index 0000000..84b85b9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/meta-reference-gpu.md @@ -0,0 +1,125 @@ +--- +orphan: true +--- + +# Meta Reference GPU Distribution + +```{toctree} +:maxdepth: 2 +:hidden: + +self +``` + +The `llamastack/distribution-meta-reference-gpu` distribution consists of the following provider configurations: + +| API | Provider(s) | +|-----|-------------| +| agents | `inline::meta-reference` | +| datasetio | `remote::huggingface`, `inline::localfs` | +| eval | `inline::meta-reference` | +| inference | `inline::meta-reference` | +| safety | `inline::llama-guard` | +| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | +| telemetry | `inline::meta-reference` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::rag-runtime`, `remote::model-context-protocol` | +| vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | + + +Note that you need access to nvidia GPUs to run this distribution. This distribution is not compatible with CPU-only machines or machines with AMD GPUs. + +### Environment Variables + +The following environment variables can be configured: + +- `LLAMA_STACK_PORT`: Port for the Llama Stack distribution server (default: `8321`) +- `INFERENCE_MODEL`: Inference model loaded into the Meta Reference server (default: `meta-llama/Llama-3.2-3B-Instruct`) +- `INFERENCE_CHECKPOINT_DIR`: Directory containing the Meta Reference model checkpoint (default: `null`) +- `SAFETY_MODEL`: Name of the safety (Llama-Guard) model to use (default: `meta-llama/Llama-Guard-3-1B`) +- `SAFETY_CHECKPOINT_DIR`: Directory containing the Llama-Guard model checkpoint (default: `null`) + + +## Prerequisite: Downloading Models + +Please use `llama model list --downloaded` to check that you have llama model checkpoints downloaded in `~/.llama` before proceeding. See [installation guide](../../references/llama_cli_reference/download_models.md) here to download the models. Run `llama model list` to see the available models to download, and `llama model download` to download the checkpoints. + +``` +$ llama model list --downloaded +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ Model โ”ƒ Size โ”ƒ Modified Time โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ Llama3.2-1B-Instruct:int4-qlora-eo8 โ”‚ 1.53 GB โ”‚ 2025-02-26 11:22:28 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-1B โ”‚ 2.31 GB โ”‚ 2025-02-18 21:48:52 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Prompt-Guard-86M โ”‚ 0.02 GB โ”‚ 2025-02-26 11:29:28 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-3B-Instruct:int4-spinquant-eo8 โ”‚ 3.69 GB โ”‚ 2025-02-26 11:37:41 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-3B โ”‚ 5.99 GB โ”‚ 2025-02-18 21:51:26 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.1-8B โ”‚ 14.97 GB โ”‚ 2025-02-16 10:36:37 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-1B-Instruct:int4-spinquant-eo8 โ”‚ 1.51 GB โ”‚ 2025-02-26 11:35:02 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama-Guard-3-1B โ”‚ 2.80 GB โ”‚ 2025-02-26 11:20:46 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama-Guard-3-1B:int4 โ”‚ 0.43 GB โ”‚ 2025-02-26 11:33:33 โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ +``` + +## Running the Distribution + +You can do this via venv or Docker which has a pre-built image. + +### Via Docker + +This method allows you to get started quickly without having to build the distribution code. + +```bash +LLAMA_STACK_PORT=8321 +docker run \ + -it \ + --pull always \ + --gpu all \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ~/.llama:/root/.llama \ + llamastack/distribution-meta-reference-gpu \ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct +``` + +If you are using Llama Stack Safety / Shield APIs, use: + +```bash +docker run \ + -it \ + --pull always \ + --gpu all \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ~/.llama:/root/.llama \ + llamastack/distribution-meta-reference-gpu \ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \ + --env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B +``` + +### Via venv + +Make sure you have done `uv pip install llama-stack` and have the Llama Stack CLI available. + +```bash +llama stack build --distro meta-reference-gpu --image-type venv +llama stack run distributions/meta-reference-gpu/run.yaml \ + --port 8321 \ + --env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct +``` + +If you are using Llama Stack Safety / Shield APIs, use: + +```bash +llama stack run distributions/meta-reference-gpu/run-with-safety.yaml \ + --port 8321 \ + --env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \ + --env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B +``` diff --git a/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/nvidia.md b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/nvidia.md new file mode 100644 index 0000000..1e52797 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/nvidia.md @@ -0,0 +1,152 @@ +--- +orphan: true +--- + +# NVIDIA Distribution + +The `llamastack/distribution-nvidia` distribution consists of the following provider configurations. + +| API | Provider(s) | +|-----|-------------| +| agents | `inline::meta-reference` | +| datasetio | `inline::localfs`, `remote::nvidia` | +| eval | `remote::nvidia` | +| files | `inline::localfs` | +| inference | `remote::nvidia` | +| post_training | `remote::nvidia` | +| safety | `remote::nvidia` | +| scoring | `inline::basic` | +| telemetry | `inline::meta-reference` | +| tool_runtime | `inline::rag-runtime` | +| vector_io | `inline::faiss` | + + +### Environment Variables + +The following environment variables can be configured: + +- `NVIDIA_API_KEY`: NVIDIA API Key (default: ``) +- `NVIDIA_APPEND_API_VERSION`: Whether to append the API version to the base_url (default: `True`) +- `NVIDIA_DATASET_NAMESPACE`: NVIDIA Dataset Namespace (default: `default`) +- `NVIDIA_PROJECT_ID`: NVIDIA Project ID (default: `test-project`) +- `NVIDIA_CUSTOMIZER_URL`: NVIDIA Customizer URL (default: `https://customizer.api.nvidia.com`) +- `NVIDIA_OUTPUT_MODEL_DIR`: NVIDIA Output Model Directory (default: `test-example-model@v1`) +- `GUARDRAILS_SERVICE_URL`: URL for the NeMo Guardrails Service (default: `http://0.0.0.0:7331`) +- `NVIDIA_GUARDRAILS_CONFIG_ID`: NVIDIA Guardrail Configuration ID (default: `self-check`) +- `NVIDIA_EVALUATOR_URL`: URL for the NeMo Evaluator Service (default: `http://0.0.0.0:7331`) +- `INFERENCE_MODEL`: Inference model (default: `Llama3.1-8B-Instruct`) +- `SAFETY_MODEL`: Name of the model to use for safety (default: `meta/llama-3.1-8b-instruct`) + + + +## Prerequisites +### NVIDIA API Keys + +Make sure you have access to a NVIDIA API Key. You can get one by visiting [https://build.nvidia.com/](https://build.nvidia.com/). Use this key for the `NVIDIA_API_KEY` environment variable. + +### Deploy NeMo Microservices Platform +The NVIDIA NeMo microservices platform supports end-to-end microservice deployment of a complete AI flywheel on your Kubernetes cluster through the NeMo Microservices Helm Chart. Please reference the [NVIDIA NeMo Microservices documentation](https://docs.nvidia.com/nemo/microservices/latest/about/index.html) for platform prerequisites and instructions to install and deploy the platform. + +## Supported Services +Each Llama Stack API corresponds to a specific NeMo microservice. The core microservices (Customizer, Evaluator, Guardrails) are exposed by the same endpoint. The platform components (Data Store) are each exposed by separate endpoints. + +### Inference: NVIDIA NIM +NVIDIA NIM is used for running inference with registered models. There are two ways to access NVIDIA NIMs: + 1. Hosted (default): Preview APIs hosted at https://integrate.api.nvidia.com (Requires an API key) + 2. Self-hosted: NVIDIA NIMs that run on your own infrastructure. + +The deployed platform includes the NIM Proxy microservice, which is the service that provides to access your NIMs (for example, to run inference on a model). Set the `NVIDIA_BASE_URL` environment variable to use your NVIDIA NIM Proxy deployment. + +### Datasetio API: NeMo Data Store +The NeMo Data Store microservice serves as the default file storage solution for the NeMo microservices platform. It exposts APIs compatible with the Hugging Face Hub client (`HfApi`), so you can use the client to interact with Data Store. The `NVIDIA_DATASETS_URL` environment variable should point to your NeMo Data Store endpoint. + +See the [NVIDIA Datasetio docs](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/remote/datasetio/nvidia/README.md) for supported features and example usage. + +### Eval API: NeMo Evaluator +The NeMo Evaluator microservice supports evaluation of LLMs. Launching an Evaluation job with NeMo Evaluator requires an Evaluation Config (an object that contains metadata needed by the job). A Llama Stack Benchmark maps to an Evaluation Config, so registering a Benchmark creates an Evaluation Config in NeMo Evaluator. The `NVIDIA_EVALUATOR_URL` environment variable should point to your NeMo Microservices endpoint. + +See the [NVIDIA Eval docs](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/remote/eval/nvidia/README.md) for supported features and example usage. + +### Post-Training API: NeMo Customizer +The NeMo Customizer microservice supports fine-tuning models. You can reference [this list of supported models](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/remote/post_training/nvidia/models.py) that can be fine-tuned using Llama Stack. The `NVIDIA_CUSTOMIZER_URL` environment variable should point to your NeMo Microservices endpoint. + +See the [NVIDIA Post-Training docs](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/remote/post_training/nvidia/README.md) for supported features and example usage. + +### Safety API: NeMo Guardrails +The NeMo Guardrails microservice sits between your application and the LLM, and adds checks and content moderation to a model. The `GUARDRAILS_SERVICE_URL` environment variable should point to your NeMo Microservices endpoint. + +See the [NVIDIA Safety docs](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/remote/safety/nvidia/README.md) for supported features and example usage. + +## Deploying models +In order to use a registered model with the Llama Stack APIs, ensure the corresponding NIM is deployed to your environment. For example, you can use the NIM Proxy microservice to deploy `meta/llama-3.2-1b-instruct`. + +Note: For improved inference speeds, we need to use NIM with `fast_outlines` guided decoding system (specified in the request body). This is the default if you deployed the platform with the NeMo Microservices Helm Chart. +```sh +# URL to NeMo NIM Proxy service +export NEMO_URL="http://nemo.test" + +curl --location "$NEMO_URL/v1/deployment/model-deployments" \ + -H 'accept: application/json' \ + -H 'Content-Type: application/json' \ + -d '{ + "name": "llama-3.2-1b-instruct", + "namespace": "meta", + "config": { + "model": "meta/llama-3.2-1b-instruct", + "nim_deployment": { + "image_name": "nvcr.io/nim/meta/llama-3.2-1b-instruct", + "image_tag": "1.8.3", + "pvc_size": "25Gi", + "gpu": 1, + "additional_envs": { + "NIM_GUIDED_DECODING_BACKEND": "fast_outlines" + } + } + } + }' +``` +This NIM deployment should take approximately 10 minutes to go live. [See the docs](https://docs.nvidia.com/nemo/microservices/latest/get-started/tutorials/deploy-nims.html) for more information on how to deploy a NIM and verify it's available for inference. + +You can also remove a deployed NIM to free up GPU resources, if needed. +```sh +export NEMO_URL="http://nemo.test" + +curl -X DELETE "$NEMO_URL/v1/deployment/model-deployments/meta/llama-3.1-8b-instruct" +``` + +## Running Llama Stack with NVIDIA + +You can do this via venv (build code), or Docker which has a pre-built image. + +### Via Docker + +This method allows you to get started quickly without having to build the distribution code. + +```bash +LLAMA_STACK_PORT=8321 +docker run \ + -it \ + --pull always \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ./run.yaml:/root/my-run.yaml \ + llamastack/distribution-nvidia \ + --config /root/my-run.yaml \ + --port $LLAMA_STACK_PORT \ + --env NVIDIA_API_KEY=$NVIDIA_API_KEY +``` + +### Via venv + +If you've set up your local development environment, you can also build the image using your local virtual environment. + +```bash +INFERENCE_MODEL=meta-llama/Llama-3.1-8B-Instruct +llama stack build --distro nvidia --image-type venv +llama stack run ./run.yaml \ + --port 8321 \ + --env NVIDIA_API_KEY=$NVIDIA_API_KEY \ + --env INFERENCE_MODEL=$INFERENCE_MODEL +``` + +## Example Notebooks +For examples of how to use the NVIDIA Distribution to run inference, fine-tune, evaluate, and run safety checks on your LLMs, you can reference the example notebooks in [docs/notebooks/nvidia](https://github.com/meta-llama/llama-stack/tree/main/docs/notebooks/nvidia). diff --git a/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/passthrough.md b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/passthrough.md new file mode 100644 index 0000000..39f076b --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/passthrough.md @@ -0,0 +1,42 @@ +--- +orphan: true +--- + +# Passthrough Distribution + +```{toctree} +:maxdepth: 2 +:hidden: + +self +``` + +The `llamastack/distribution-passthrough` distribution consists of the following provider configurations. + +| API | Provider(s) | +|-----|-------------| +| agents | `inline::meta-reference` | +| datasetio | `remote::huggingface`, `inline::localfs` | +| eval | `inline::meta-reference` | +| inference | `remote::passthrough`, `inline::sentence-transformers` | +| safety | `inline::llama-guard` | +| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | +| telemetry | `inline::meta-reference` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `remote::wolfram-alpha`, `inline::rag-runtime`, `remote::model-context-protocol` | +| vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | + + +### Environment Variables + +The following environment variables can be configured: + +- `LLAMA_STACK_PORT`: Port for the Llama Stack distribution server (default: `8321`) +- `PASSTHROUGH_API_KEY`: Passthrough API Key (default: ``) +- `PASSTHROUGH_URL`: Passthrough URL (default: ``) + +### Models + +The following models are available by default: + +- `llama3.1-8b-instruct ` +- `llama3.2-11b-vision-instruct ` diff --git a/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/starter.md b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/starter.md new file mode 100644 index 0000000..faa82bc --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/self_hosted_distro/starter.md @@ -0,0 +1,232 @@ +--- +orphan: true +--- + +# Starter Distribution + +```{toctree} +:maxdepth: 2 +:hidden: + +self +``` + +The `llamastack/distribution-starter` distribution is a comprehensive, multi-provider distribution that includes most of the available inference providers in Llama Stack. It's designed to be a one-stop solution for developers who want to experiment with different AI providers without having to configure each one individually. + +## Provider Composition + +The starter distribution consists of the following provider configurations: + +| API | Provider(s) | +|-----|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| +| agents | `inline::meta-reference` | +| datasetio | `remote::huggingface`, `inline::localfs` | +| eval | `inline::meta-reference` | +| files | `inline::localfs` | +| inference | `remote::openai`, `remote::fireworks`, `remote::together`, `remote::ollama`, `remote::anthropic`, `remote::gemini`, `remote::groq`, `remote::sambanova`, `remote::vllm`, `remote::tgi`, `remote::cerebras`, `remote::llama-openai-compat`, `remote::nvidia`, `remote::hf::serverless`, `remote::hf::endpoint`, `inline::sentence-transformers` | +| safety | `inline::llama-guard` | +| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | +| telemetry | `inline::meta-reference` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::rag-runtime`, `remote::model-context-protocol` | +| vector_io | `inline::faiss`, `inline::sqlite-vec`, `inline::milvus`, `remote::chromadb`, `remote::pgvector` | + +## Inference Providers + +The starter distribution includes a comprehensive set of inference providers: + +### Hosted Providers +- **[OpenAI](https://openai.com/api/)**: GPT-4, GPT-3.5, O1, O3, O4 models and text embeddings - + provider ID: `openai` - reference documentation: [openai](../../providers/inference/remote_openai) +- **[Fireworks](https://fireworks.ai/)**: Llama 3.1, 3.2, 3.3, 4 Scout, 4 Maverick models and + embeddings - provider ID: `fireworks` - reference documentation: [fireworks](../../providers/inference/remote_fireworks) +- **[Together](https://together.ai/)**: Llama 3.1, 3.2, 3.3, 4 Scout, 4 Maverick models and + embeddings - provider ID: `together` - reference documentation: [together](../../providers/inference/remote_together) +- **[Anthropic](https://www.anthropic.com/)**: Claude 3.5 Sonnet, Claude 3.7 Sonnet, Claude 3.5 Haiku, and Voyage embeddings - provider ID: `anthropic` - reference documentation: [anthropic](../../providers/inference/remote_anthropic) +- **[Gemini](https://gemini.google.com/)**: Gemini 1.5, 2.0, 2.5 models and text embeddings - provider ID: `gemini` - reference documentation: [gemini](../../providers/inference/remote_gemini) +- **[Groq](https://groq.com/)**: Fast Llama models (3.1, 3.2, 3.3, 4 Scout, 4 Maverick) - provider ID: `groq` - reference documentation: [groq](../../providers/inference/remote_groq) +- **[SambaNova](https://www.sambanova.ai/)**: Llama 3.1, 3.2, 3.3, 4 Scout, 4 Maverick models - provider ID: `sambanova` - reference documentation: [sambanova](../../providers/inference/remote_sambanova) +- **[Cerebras](https://www.cerebras.ai/)**: Cerebras AI models - provider ID: `cerebras` - reference documentation: [cerebras](../../providers/inference/remote_cerebras) +- **[NVIDIA](https://www.nvidia.com/)**: NVIDIA NIM - provider ID: `nvidia` - reference documentation: [nvidia](../../providers/inference/remote_nvidia) +- **[HuggingFace](https://huggingface.co/)**: Serverless and endpoint models - provider ID: `hf::serverless` and `hf::endpoint` - reference documentation: [huggingface-serverless](../../providers/inference/remote_hf_serverless) and [huggingface-endpoint](../../providers/inference/remote_hf_endpoint) +- **[Bedrock](https://aws.amazon.com/bedrock/)**: AWS Bedrock models - provider ID: `bedrock` - reference documentation: [bedrock](../../providers/inference/remote_bedrock) + +### Local/Remote Providers +- **[Ollama](https://ollama.ai/)**: Local Ollama models - provider ID: `ollama` - reference documentation: [ollama](../../providers/inference/remote_ollama) +- **[vLLM](https://docs.vllm.ai/en/latest/)**: Local or remote vLLM server - provider ID: `vllm` - reference documentation: [vllm](../../providers/inference/remote_vllm) +- **[TGI](https://github.com/huggingface/text-generation-inference)**: Text Generation Inference server - Dell Enterprise Hub's custom TGI container too (use `DEH_URL`) - provider ID: `tgi` - reference documentation: [tgi](../../providers/inference/remote_tgi) +- **[Sentence Transformers](https://www.sbert.net/)**: Local embedding models - provider ID: `sentence-transformers` - reference documentation: [sentence-transformers](../../providers/inference/inline_sentence-transformers) + +All providers are disabled by default. So you need to enable them by setting the environment variables. + +## Vector IO + +The starter distribution includes a comprehensive set of vector IO providers: + +- **[FAISS](https://github.com/facebookresearch/faiss)**: Local FAISS vector store - enabled by + default - provider ID: `faiss` +- **[SQLite](https://www.sqlite.org/index.html)**: Local SQLite vector store - disabled by default - provider ID: `sqlite-vec` +- **[ChromaDB](https://www.trychroma.com/)**: Remote ChromaDB vector store - disabled by default - provider ID: `chromadb` +- **[PGVector](https://github.com/pgvector/pgvector)**: PostgreSQL vector store - disabled by default - provider ID: `pgvector` +- **[Milvus](https://milvus.io/)**: Milvus vector store - disabled by default - provider ID: `milvus` + +## Environment Variables + +The following environment variables can be configured: + +### Server Configuration +- `LLAMA_STACK_PORT`: Port for the Llama Stack distribution server (default: `8321`) + +### API Keys for Hosted Providers +- `OPENAI_API_KEY`: OpenAI API key +- `FIREWORKS_API_KEY`: Fireworks API key +- `TOGETHER_API_KEY`: Together API key +- `ANTHROPIC_API_KEY`: Anthropic API key +- `GEMINI_API_KEY`: Google Gemini API key +- `GROQ_API_KEY`: Groq API key +- `SAMBANOVA_API_KEY`: SambaNova API key +- `CEREBRAS_API_KEY`: Cerebras API key +- `LLAMA_API_KEY`: Llama API key +- `NVIDIA_API_KEY`: NVIDIA API key +- `HF_API_TOKEN`: HuggingFace API token + +### Local Provider Configuration +- `OLLAMA_URL`: Ollama server URL (default: `http://localhost:11434`) +- `VLLM_URL`: vLLM server URL (default: `http://localhost:8000/v1`) +- `VLLM_MAX_TOKENS`: vLLM max tokens (default: `4096`) +- `VLLM_API_TOKEN`: vLLM API token (default: `fake`) +- `VLLM_TLS_VERIFY`: vLLM TLS verification (default: `true`) +- `TGI_URL`: TGI server URL + +### Model Configuration +- `INFERENCE_MODEL`: HuggingFace model for serverless inference +- `INFERENCE_ENDPOINT_NAME`: HuggingFace endpoint name + +### Vector Database Configuration +- `SQLITE_STORE_DIR`: SQLite store directory (default: `~/.llama/distributions/starter`) +- `ENABLE_SQLITE_VEC`: Enable SQLite vector provider +- `ENABLE_CHROMADB`: Enable ChromaDB provider +- `ENABLE_PGVECTOR`: Enable PGVector provider +- `CHROMADB_URL`: ChromaDB server URL +- `PGVECTOR_HOST`: PGVector host (default: `localhost`) +- `PGVECTOR_PORT`: PGVector port (default: `5432`) +- `PGVECTOR_DB`: PGVector database name +- `PGVECTOR_USER`: PGVector username +- `PGVECTOR_PASSWORD`: PGVector password + +### Tool Configuration +- `BRAVE_SEARCH_API_KEY`: Brave Search API key +- `TAVILY_SEARCH_API_KEY`: Tavily Search API key + +### Telemetry Configuration +- `OTEL_SERVICE_NAME`: OpenTelemetry service name +- `TELEMETRY_SINKS`: Telemetry sinks (default: `console,sqlite`) + +## Enabling Providers + +You can enable specific providers by setting appropriate environment variables. For example, + +```bash +# self-hosted +export OLLAMA_URL=http://localhost:11434 # enables the Ollama inference provider +export VLLM_URL=http://localhost:8000/v1 # enables the vLLM inference provider +export TGI_URL=http://localhost:8000/v1 # enables the TGI inference provider + +# cloud-hosted requiring API key configuration on the server +export CEREBRAS_API_KEY=your_cerebras_api_key # enables the Cerebras inference provider +export NVIDIA_API_KEY=your_nvidia_api_key # enables the NVIDIA inference provider + +# vector providers +export MILVUS_URL=http://localhost:19530 # enables the Milvus vector provider +export CHROMADB_URL=http://localhost:8000/v1 # enables the ChromaDB vector provider +export PGVECTOR_DB=llama_stack_db # enables the PGVector vector provider +``` + +This distribution comes with a default "llama-guard" shield that can be enabled by setting the `SAFETY_MODEL` environment variable to point to an appropriate Llama Guard model id. Use `llama-stack-client models list` to see the list of available models. + +## Running the Distribution + +You can run the starter distribution via Docker or venv. + +### Via Docker + +This method allows you to get started quickly without having to build the distribution code. + +```bash +LLAMA_STACK_PORT=8321 +docker run \ + -it \ + --pull always \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -e OPENAI_API_KEY=your_openai_key \ + -e FIREWORKS_API_KEY=your_fireworks_key \ + -e TOGETHER_API_KEY=your_together_key \ + llamastack/distribution-starter \ + --port $LLAMA_STACK_PORT +``` + +### Via venv + +Ensure you have configured the starter distribution using the environment variables explained above. + +```bash +uv run --with llama-stack llama stack build --distro starter --image-type venv --run +``` + +## Example Usage + +Once the distribution is running, you can use any of the available models. Here are some examples: + +### Using OpenAI Models +```bash +llama-stack-client --endpoint http://localhost:8321 \ +inference chat-completion \ +--model-id openai/gpt-4o \ +--message "Hello, how are you?" +``` + +### Using Fireworks Models +```bash +llama-stack-client --endpoint http://localhost:8321 \ +inference chat-completion \ +--model-id fireworks/meta-llama/Llama-3.2-3B-Instruct \ +--message "Write a short story about a robot." +``` + +### Using Local Ollama Models +```bash +# First, make sure Ollama is running and you have a model +ollama run llama3.2:3b + +# Then use it through Llama Stack +export OLLAMA_INFERENCE_MODEL=llama3.2:3b +llama-stack-client --endpoint http://localhost:8321 \ +inference chat-completion \ +--model-id ollama/llama3.2:3b \ +--message "Explain quantum computing in simple terms." +``` + +## Storage + +The starter distribution uses SQLite for local storage of various components: + +- **Metadata store**: `~/.llama/distributions/starter/registry.db` +- **Inference store**: `~/.llama/distributions/starter/inference_store.db` +- **FAISS store**: `~/.llama/distributions/starter/faiss_store.db` +- **SQLite vector store**: `~/.llama/distributions/starter/sqlite_vec.db` +- **Files metadata**: `~/.llama/distributions/starter/files_metadata.db` +- **Agents store**: `~/.llama/distributions/starter/agents_store.db` +- **Responses store**: `~/.llama/distributions/starter/responses_store.db` +- **Trace store**: `~/.llama/distributions/starter/trace_store.db` +- **Evaluation store**: `~/.llama/distributions/starter/meta_reference_eval.db` +- **Dataset I/O stores**: Various HuggingFace and local filesystem stores + +## Benefits of the Starter Distribution + +1. **Comprehensive Coverage**: Includes most popular AI providers in one distribution +2. **Flexible Configuration**: Easy to enable/disable providers based on your needs +3. **No Local GPU Required**: Most providers are cloud-based, making it accessible to developers without high-end hardware +4. **Easy Migration**: Start with hosted providers and gradually move to local ones as needed +5. **Production Ready**: Includes safety, evaluation, and telemetry components +6. **Tool Integration**: Comes with web search, RAG, and model context protocol tools + +The starter distribution is ideal for developers who want to experiment with different AI providers, build prototypes quickly, or create applications that can work with multiple AI backends. diff --git a/versioned_docs/version-v0.2.23/distributions/starting_llama_stack_server.mdx b/versioned_docs/version-v0.2.23/distributions/starting_llama_stack_server.mdx new file mode 100644 index 0000000..0260692 --- /dev/null +++ b/versioned_docs/version-v0.2.23/distributions/starting_llama_stack_server.mdx @@ -0,0 +1,32 @@ +--- +title: Starting a Llama Stack Server +description: Different ways to run Llama Stack servers - as library, container, or Kubernetes deployment +sidebar_label: Starting Llama Stack Server +sidebar_position: 7 +--- + +# Starting a Llama Stack Server + +You can run a Llama Stack server in one of the following ways: + +## As a Library: + +This is the simplest way to get started. Using Llama Stack as a library means you do not need to start a server. This is especially useful when you are not running inference locally and relying on an external inference service (eg. fireworks, together, groq, etc.) See [Using Llama Stack as a Library](importing_as_library) + + +## Container: + +Another simple way to start interacting with Llama Stack is to just spin up a container (via Docker or Podman) which is pre-built with all the providers you need. We provide a number of pre-built images so you can start a Llama Stack server instantly. You can also build your own custom container. Which distribution to choose depends on the hardware you have. See [Selection of a Distribution](./list_of_distributions) for more details. + +## Kubernetes: + +If you have built a container image and want to deploy it in a Kubernetes cluster instead of starting the Llama Stack server locally. See [Kubernetes Deployment Guide](../deploying/kubernetes_deployment) for more details. + + +```{toctree} +:maxdepth: 1 +:hidden: + +importing_as_library +configuration +``` diff --git a/versioned_docs/version-v0.2.23/getting_started/demo_script.py b/versioned_docs/version-v0.2.23/getting_started/demo_script.py new file mode 100644 index 0000000..2ea6773 --- /dev/null +++ b/versioned_docs/version-v0.2.23/getting_started/demo_script.py @@ -0,0 +1,68 @@ +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +from llama_stack_client import Agent, AgentEventLogger, RAGDocument, LlamaStackClient + +vector_db_id = "my_demo_vector_db" +client = LlamaStackClient(base_url="http://localhost:8321") + +models = client.models.list() + +# Select the first LLM and first embedding models +model_id = next(m for m in models if m.model_type == "llm").identifier +embedding_model_id = ( + em := next(m for m in models if m.model_type == "embedding") +).identifier +embedding_dimension = em.metadata["embedding_dimension"] + +vector_db = client.vector_dbs.register( + vector_db_id=vector_db_id, + embedding_model=embedding_model_id, + embedding_dimension=embedding_dimension, + provider_id="faiss", +) +vector_db_id = vector_db.identifier +source = "https://www.paulgraham.com/greatwork.html" +print("rag_tool> Ingesting document:", source) +document = RAGDocument( + document_id="document_1", + content=source, + mime_type="text/html", + metadata={}, +) +client.tool_runtime.rag_tool.insert( + documents=[document], + vector_db_id=vector_db_id, + chunk_size_in_tokens=100, +) +agent = Agent( + client, + model=model_id, + instructions="You are a helpful assistant", + tools=[ + { + "name": "builtin::rag/knowledge_search", + "args": {"vector_db_ids": [vector_db_id]}, + } + ], +) + +prompt = "How do you do great work?" +print("prompt>", prompt) + +use_stream = True +response = agent.create_turn( + messages=[{"role": "user", "content": prompt}], + session_id=agent.create_session("rag_session"), + stream=use_stream, +) + +# Only call `AgentEventLogger().log(response)` for streaming responses. +if use_stream: + for log in AgentEventLogger().log(response): + log.print() +else: + print(response) diff --git a/versioned_docs/version-v0.2.23/getting_started/detailed_tutorial.mdx b/versioned_docs/version-v0.2.23/getting_started/detailed_tutorial.mdx new file mode 100644 index 0000000..33786ac --- /dev/null +++ b/versioned_docs/version-v0.2.23/getting_started/detailed_tutorial.mdx @@ -0,0 +1,541 @@ +--- +title: Detailed Tutorial +description: Complete guide to using Llama Stack server and client SDK to build AI agents +sidebar_label: Detailed Tutorial +sidebar_position: 3 +--- + +import Tabs from '@theme/Tabs'; +import TabItem from '@theme/TabItem'; + +## Detailed Tutorial + +In this guide, we'll walk through how you can use the Llama Stack (server and client SDK) to test a simple agent. +A Llama Stack agent is a simple integrated system that can perform tasks by combining a Llama model for reasoning with +tools (e.g., RAG, web search, code execution, etc.) for taking actions. +In Llama Stack, we provide a server exposing multiple APIs. These APIs are backed by implementations from different providers. + +Llama Stack is a stateful service with REST APIs to support seamless transition of AI applications across different environments. The server can be run in a variety of ways, including as a standalone binary, Docker container, or hosted service. You can build and test using a local server first and deploy to a hosted endpoint for production. + +In this guide, we'll walk through how to build a RAG agent locally using Llama Stack with [Ollama](https://ollama.com/) +as the inference [provider](/docs/providers/inference/) for a Llama Model. + +### Step 1: Installation and Setup + +Install Ollama by following the instructions on the [Ollama website](https://ollama.com/download), then +download Llama 3.2 3B model, and then start the Ollama service. +```bash +ollama pull llama3.2:3b +ollama run llama3.2:3b --keepalive 60m +``` + +Install [uv](https://docs.astral.sh/uv/) to setup your virtual environment + + + +Use `curl` to download the script and execute it with `sh`: +```console +curl -LsSf https://astral.sh/uv/install.sh | sh +``` + + +Use `irm` to download the script and execute it with `iex`: + +```console +powershell -ExecutionPolicy ByPass -c "irm https://astral.sh/uv/install.ps1 | iex" +``` + + + +Setup your virtual environment. + +```bash +uv sync --python 3.12 +source .venv/bin/activate +``` +### Step 2: Run Llama Stack +Llama Stack is a server that exposes multiple APIs, you connect with it using the Llama Stack client SDK. + + + +You can use Python to build and run the Llama Stack server, which is useful for testing and development. + +Llama Stack uses a [YAML configuration file](../distributions/configuration) to specify the stack setup, +which defines the providers and their settings. The generated configuration serves as a starting point that you can [customize for your specific needs](../distributions/customizing_run_yaml). +Now let's build and run the Llama Stack config for Ollama. +We use `starter` as template. By default all providers are disabled, this requires enable ollama by passing environment variables. + +```bash +llama stack build --distro starter --image-type venv --run +``` + + +You can use a container image to run the Llama Stack server. We provide several container images for the server +component that works with different inference providers out of the box. For this guide, we will use +`llamastack/distribution-starter` as the container image. If you'd like to build your own image or customize the +configurations, please check out [this guide](../distributions/building_distro). +First lets setup some environment variables and create a local directory to mount into the containerโ€™s file system. +```bash +export LLAMA_STACK_PORT=8321 +mkdir -p ~/.llama +``` +Then start the server using the container tool of your choice. For example, if you are running Docker you can use the +following command: +```bash +docker run -it \ + --pull always \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ~/.llama:/root/.llama \ + llamastack/distribution-starter \ + --port $LLAMA_STACK_PORT \ + --env OLLAMA_URL=http://host.docker.internal:11434 +``` +Note to start the container with Podman, you can do the same but replace `docker` at the start of the command with +`podman`. If you are using `podman` older than `4.7.0`, please also replace `host.docker.internal` in the `OLLAMA_URL` +with `host.containers.internal`. + +The configuration YAML for the Ollama distribution is available at `distributions/ollama/run.yaml`. + +:::tip +Docker containers run in their own isolated network namespaces on Linux. To allow the container to communicate with services running on the host via `localhost`, you need `--network=host`. This makes the container use the host's network directly so it can connect to Ollama running on `localhost:11434`. + +Linux users having issues running the above command should instead try the following: +```bash +docker run -it \ + --pull always \ + -p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \ + -v ~/.llama:/root/.llama \ + --network=host \ + llamastack/distribution-starter \ + --port $LLAMA_STACK_PORT \ + --env OLLAMA_URL=http://localhost:11434 +``` +::: +You will see output like below: +``` +INFO: Application startup complete. +INFO: Uvicorn running on http://['::', '0.0.0.0']:8321 (Press CTRL+C to quit) +``` + +Now you can use the Llama Stack client to run inference and build agents! + +You can reuse the server setup or use the [Llama Stack Client](https://github.com/meta-llama/llama-stack-client-python/). +Note that the client package is already included in the `llama-stack` package. + + + +### Step 3: Run Client CLI + +Open a new terminal and navigate to the same directory you started the server from. Then set up a new or activate your +existing server virtual environment. + + + +```bash +# The client is included in the llama-stack package so we just activate the server venv +source .venv/bin/activate +``` + + +```bash +uv venv client --python 3.12 +source client/bin/activate +pip install llama-stack-client +``` + + + +Now let's use the `llama-stack-client` [CLI](../references/llama_stack_client_cli_reference) to check the +connectivity to the server. + +```bash +llama-stack-client configure --endpoint http://localhost:8321 --api-key none +``` +You will see the below: +``` +Done! You can now use the Llama Stack Client CLI with endpoint http://localhost:8321 +``` + +List the models +```bash +llama-stack-client models list +Available Models + +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ model_type โ”ƒ identifier โ”ƒ provider_resource_id โ”ƒ metadata โ”ƒ provider_id โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ embedding โ”‚ ollama/all-minilm:l6-v2 โ”‚ all-minilm:l6-v2 โ”‚ {'embedding_dimension': 384.0} โ”‚ ollama โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ ... โ”‚ ... โ”‚ ... โ”‚ โ”‚ ... โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ llm โ”‚ ollama/Llama-3.2:3b โ”‚ llama3.2:3b โ”‚ โ”‚ ollama โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ + +``` +You can test basic Llama inference completion using the CLI. + +```bash +llama-stack-client inference chat-completion --model-id "ollama/llama3.2:3b" --message "tell me a joke" + +``` +Sample output: +```python +OpenAIChatCompletion( + id="chatcmpl-08d7b2be-40f3-47ed-8f16-a6f29f2436af", + choices=[ + OpenAIChatCompletionChoice( + finish_reason="stop", + index=0, + message=OpenAIChatCompletionChoiceMessageOpenAIAssistantMessageParam( + role="assistant", + content="Why couldn't the bicycle stand up by itself?\n\nBecause it was two-tired.", + name=None, + tool_calls=None, + refusal=None, + annotations=None, + audio=None, + function_call=None, + ), + logprobs=None, + ) + ], + created=1751725254, + model="llama3.2:3b", + object="chat.completion", + service_tier=None, + system_fingerprint="fp_ollama", + usage={ + "completion_tokens": 18, + "prompt_tokens": 29, + "total_tokens": 47, + "completion_tokens_details": None, + "prompt_tokens_details": None, + }, +) +``` + +### Step 4: Run the Demos + +Note that these demos show the [Python Client SDK](../references/python_sdk_reference/). +Other SDKs are also available, please refer to the [Client SDK](/docs/) list for the complete options. + + + +Now you can run inference using the Llama Stack client SDK. + +#### i. Create the Script + +Create a file `inference.py` and add the following code: +```python +from llama_stack_client import LlamaStackClient + +client = LlamaStackClient(base_url="http://localhost:8321") + +# List available models +models = client.models.list() + +# Select the first LLM +llm = next(m for m in models if m.model_type == "llm" and m.provider_id == "ollama") +model_id = llm.identifier + +print("Model:", model_id) + +response = client.chat.completions.create( + model=model_id, + messages=[ + {"role": "system", "content": "You are a helpful assistant."}, + {"role": "user", "content": "Write a haiku about coding"}, + ], +) +print(response) +``` + +#### ii. Run the Script +Let's run the script using `uv` +```bash +uv run python inference.py +``` +Which will output: +``` +Model: ollama/llama3.2:3b +OpenAIChatCompletion(id='chatcmpl-30cd0f28-a2ad-4b6d-934b-13707fc60ebf', choices=[OpenAIChatCompletionChoice(finish_reason='stop', index=0, message=OpenAIChatCompletionChoiceMessageOpenAIAssistantMessageParam(role='assistant', content="Lines of code unfold\nAlgorithms dance with ease\nLogic's gentle kiss", name=None, tool_calls=None, refusal=None, annotations=None, audio=None, function_call=None), logprobs=None)], created=1751732480, model='llama3.2:3b', object='chat.completion', service_tier=None, system_fingerprint='fp_ollama', usage={'completion_tokens': 16, 'prompt_tokens': 37, 'total_tokens': 53, 'completion_tokens_details': None, 'prompt_tokens_details': None}) +``` + + +Next we can move beyond simple inference and build an agent that can perform tasks using the Llama Stack server. +#### i. Create the Script +Create a file `agent.py` and add the following code: + +```python +from llama_stack_client import LlamaStackClient +from llama_stack_client import Agent, AgentEventLogger +from rich.pretty import pprint +import uuid + +client = LlamaStackClient(base_url=f"http://localhost:8321") + +models = client.models.list() +llm = next(m for m in models if m.model_type == "llm" and m.provider_id == "ollama") +model_id = llm.identifier + +agent = Agent(client, model=model_id, instructions="You are a helpful assistant.") + +s_id = agent.create_session(session_name=f"s{uuid.uuid4().hex}") + +print("Non-streaming ...") +response = agent.create_turn( + messages=[{"role": "user", "content": "Who are you?"}], + session_id=s_id, + stream=False, +) +print("agent>", response.output_message.content) + +print("Streaming ...") +stream = agent.create_turn( + messages=[{"role": "user", "content": "Who are you?"}], session_id=s_id, stream=True +) +for event in stream: + pprint(event) + +print("Streaming with print helper...") +stream = agent.create_turn( + messages=[{"role": "user", "content": "Who are you?"}], session_id=s_id, stream=True +) +for event in AgentEventLogger().log(stream): + event.print() +``` +### ii. Run the Script +Let's run the script using `uv` +```bash +uv run python agent.py +``` + +```{dropdown} ๐Ÿ‘‹ Click here to see the sample output + Non-streaming ... + agent> I'm an artificial intelligence designed to assist and communicate with users like you. I don't have a personal identity, but I can provide information, answer questions, and help with tasks to the best of my abilities. + + I'm a large language model, which means I've been trained on a massive dataset of text from various sources, allowing me to understand and respond to a wide range of topics and questions. My purpose is to provide helpful and accurate information, and I'm constantly learning and improving my responses based on the interactions I have with users like you. + + I can help with: + + * Answering questions on various subjects + * Providing definitions and explanations + * Offering suggestions and ideas + * Assisting with language-related tasks, such as proofreading and editing + * Generating text and content + * And more! + + Feel free to ask me anything, and I'll do my best to help! + Streaming ... + AgentTurnResponseStreamChunk( + โ”‚ event=TurnResponseEvent( + โ”‚ โ”‚ payload=AgentTurnResponseStepStartPayload( + โ”‚ โ”‚ โ”‚ event_type='step_start', + โ”‚ โ”‚ โ”‚ step_id='69831607-fa75-424a-949b-e2049e3129d1', + โ”‚ โ”‚ โ”‚ step_type='inference', + โ”‚ โ”‚ โ”‚ metadata={} + โ”‚ โ”‚ ) + โ”‚ ) + ) + AgentTurnResponseStreamChunk( + โ”‚ event=TurnResponseEvent( + โ”‚ โ”‚ payload=AgentTurnResponseStepProgressPayload( + โ”‚ โ”‚ โ”‚ delta=TextDelta(text='As', type='text'), + โ”‚ โ”‚ โ”‚ event_type='step_progress', + โ”‚ โ”‚ โ”‚ step_id='69831607-fa75-424a-949b-e2049e3129d1', + โ”‚ โ”‚ โ”‚ step_type='inference' + โ”‚ โ”‚ ) + โ”‚ ) + ) + AgentTurnResponseStreamChunk( + โ”‚ event=TurnResponseEvent( + โ”‚ โ”‚ payload=AgentTurnResponseStepProgressPayload( + โ”‚ โ”‚ โ”‚ delta=TextDelta(text=' a', type='text'), + โ”‚ โ”‚ โ”‚ event_type='step_progress', + โ”‚ โ”‚ โ”‚ step_id='69831607-fa75-424a-949b-e2049e3129d1', + โ”‚ โ”‚ โ”‚ step_type='inference' + โ”‚ โ”‚ ) + โ”‚ ) + ) + ... + AgentTurnResponseStreamChunk( + โ”‚ event=TurnResponseEvent( + โ”‚ โ”‚ payload=AgentTurnResponseStepCompletePayload( + โ”‚ โ”‚ โ”‚ event_type='step_complete', + โ”‚ โ”‚ โ”‚ step_details=InferenceStep( + โ”‚ โ”‚ โ”‚ โ”‚ api_model_response=CompletionMessage( + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ content='As a conversational AI, I don\'t have a personal identity in the classical sense. I exist as a program running on computer servers, designed to process and respond to text-based inputs.\n\nI\'m an instance of a type of artificial intelligence called a "language model," which is trained on vast amounts of text data to generate human-like responses. My primary function is to understand and respond to natural language inputs, like our conversation right now.\n\nThink of me as a virtual assistant, a chatbot, or a conversational interface โ€“ I\'m here to provide information, answer questions, and engage in conversation to the best of my abilities. I don\'t have feelings, emotions, or consciousness like humans do, but I\'m designed to simulate human-like interactions to make our conversations feel more natural and helpful.\n\nSo, that\'s me in a nutshell! What can I help you with today?', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ role='assistant', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ stop_reason='end_of_turn', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ tool_calls=[] + โ”‚ โ”‚ โ”‚ โ”‚ ), + โ”‚ โ”‚ โ”‚ โ”‚ step_id='69831607-fa75-424a-949b-e2049e3129d1', + โ”‚ โ”‚ โ”‚ โ”‚ step_type='inference', + โ”‚ โ”‚ โ”‚ โ”‚ turn_id='8b360202-f7cb-4786-baa9-166a1b46e2ca', + โ”‚ โ”‚ โ”‚ โ”‚ completed_at=datetime.datetime(2025, 4, 3, 1, 15, 21, 716174, tzinfo=TzInfo(UTC)), + โ”‚ โ”‚ โ”‚ โ”‚ started_at=datetime.datetime(2025, 4, 3, 1, 15, 14, 28823, tzinfo=TzInfo(UTC)) + โ”‚ โ”‚ โ”‚ ), + โ”‚ โ”‚ โ”‚ step_id='69831607-fa75-424a-949b-e2049e3129d1', + โ”‚ โ”‚ โ”‚ step_type='inference' + โ”‚ โ”‚ ) + โ”‚ ) + ) + AgentTurnResponseStreamChunk( + โ”‚ event=TurnResponseEvent( + โ”‚ โ”‚ payload=AgentTurnResponseTurnCompletePayload( + โ”‚ โ”‚ โ”‚ event_type='turn_complete', + โ”‚ โ”‚ โ”‚ turn=Turn( + โ”‚ โ”‚ โ”‚ โ”‚ input_messages=[UserMessage(content='Who are you?', role='user', context=None)], + โ”‚ โ”‚ โ”‚ โ”‚ output_message=CompletionMessage( + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ content='As a conversational AI, I don\'t have a personal identity in the classical sense. I exist as a program running on computer servers, designed to process and respond to text-based inputs.\n\nI\'m an instance of a type of artificial intelligence called a "language model," which is trained on vast amounts of text data to generate human-like responses. My primary function is to understand and respond to natural language inputs, like our conversation right now.\n\nThink of me as a virtual assistant, a chatbot, or a conversational interface โ€“ I\'m here to provide information, answer questions, and engage in conversation to the best of my abilities. I don\'t have feelings, emotions, or consciousness like humans do, but I\'m designed to simulate human-like interactions to make our conversations feel more natural and helpful.\n\nSo, that\'s me in a nutshell! What can I help you with today?', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ role='assistant', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ stop_reason='end_of_turn', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ tool_calls=[] + โ”‚ โ”‚ โ”‚ โ”‚ ), + โ”‚ โ”‚ โ”‚ โ”‚ session_id='abd4afea-4324-43f4-9513-cfe3970d92e8', + โ”‚ โ”‚ โ”‚ โ”‚ started_at=datetime.datetime(2025, 4, 3, 1, 15, 14, 28722, tzinfo=TzInfo(UTC)), + โ”‚ โ”‚ โ”‚ โ”‚ steps=[ + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ InferenceStep( + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ api_model_response=CompletionMessage( + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ content='As a conversational AI, I don\'t have a personal identity in the classical sense. I exist as a program running on computer servers, designed to process and respond to text-based inputs.\n\nI\'m an instance of a type of artificial intelligence called a "language model," which is trained on vast amounts of text data to generate human-like responses. My primary function is to understand and respond to natural language inputs, like our conversation right now.\n\nThink of me as a virtual assistant, a chatbot, or a conversational interface โ€“ I\'m here to provide information, answer questions, and engage in conversation to the best of my abilities. I don\'t have feelings, emotions, or consciousness like humans do, but I\'m designed to simulate human-like interactions to make our conversations feel more natural and helpful.\n\nSo, that\'s me in a nutshell! What can I help you with today?', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ role='assistant', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ stop_reason='end_of_turn', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ tool_calls=[] + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ ), + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ step_id='69831607-fa75-424a-949b-e2049e3129d1', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ step_type='inference', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ turn_id='8b360202-f7cb-4786-baa9-166a1b46e2ca', + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ completed_at=datetime.datetime(2025, 4, 3, 1, 15, 21, 716174, tzinfo=TzInfo(UTC)), + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ started_at=datetime.datetime(2025, 4, 3, 1, 15, 14, 28823, tzinfo=TzInfo(UTC)) + โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ ) + โ”‚ โ”‚ โ”‚ โ”‚ ], + โ”‚ โ”‚ โ”‚ โ”‚ turn_id='8b360202-f7cb-4786-baa9-166a1b46e2ca', + โ”‚ โ”‚ โ”‚ โ”‚ completed_at=datetime.datetime(2025, 4, 3, 1, 15, 21, 727364, tzinfo=TzInfo(UTC)), + โ”‚ โ”‚ โ”‚ โ”‚ output_attachments=[] + โ”‚ โ”‚ โ”‚ ) + โ”‚ โ”‚ ) + โ”‚ ) + ) + + + Streaming with print helper... + inference> Dรฉjร  vu! You're asking me again! + + As I mentioned earlier, I'm a computer program designed to simulate conversation and answer questions. I don't have a personal identity or consciousness like a human would. I exist solely as a digital entity, running on computer servers and responding to inputs from users like you. + + I'm a type of artificial intelligence (AI) called a large language model, which means I've been trained on a massive dataset of text from various sources. This training allows me to understand and respond to a wide range of questions and topics. + + My purpose is to provide helpful and accurate information, answer questions, and assist users like you with tasks and conversations. I don't have personal preferences, emotions, or opinions like humans do. My goal is to be informative, neutral, and respectful in my responses. + + So, that's me in a nutshell! +``` + + + +For our last demo, we can build a RAG agent that can answer questions about the Torchtune project using the documents +in a vector database. +#### i. Create the Script +Create a file `rag_agent.py` and add the following code: + +```python +from llama_stack_client import LlamaStackClient +from llama_stack_client import Agent, AgentEventLogger +from llama_stack_client.types import Document +import uuid + +client = LlamaStackClient(base_url="http://localhost:8321") + +# Create a vector database instance +embed_lm = next(m for m in client.models.list() if m.model_type == "embedding") +embedding_model = embed_lm.identifier +vector_db_id = f"v{uuid.uuid4().hex}" +# The VectorDB API is deprecated; the server now returns its own authoritative ID. +# We capture the correct ID from the response's .identifier attribute. +vector_db_id = client.vector_dbs.register( + vector_db_id=vector_db_id, + embedding_model=embedding_model, +).identifier + +# Create Documents +urls = [ + "memory_optimizations.rst", + "chat.rst", + "llama3.rst", + "qat_finetune.rst", + "lora_finetune.rst", +] +documents = [ + Document( + document_id=f"num-{i}", + content=f"https://raw.githubusercontent.com/pytorch/torchtune/main/docs/source/tutorials/{url}", + mime_type="text/plain", + metadata={}, + ) + for i, url in enumerate(urls) +] + +# Insert documents +client.tool_runtime.rag_tool.insert( + documents=documents, + vector_db_id=vector_db_id, + chunk_size_in_tokens=512, +) + +# Get the model being served +llm = next( + m + for m in client.models.list() + if m.model_type == "llm" and m.provider_id == "ollama" +) +model = llm.identifier + +# Create the RAG agent +rag_agent = Agent( + client, + model=model, + instructions="You are a helpful assistant. Use the RAG tool to answer questions as needed.", + tools=[ + { + "name": "builtin::rag/knowledge_search", + "args": {"vector_db_ids": [vector_db_id]}, + } + ], +) + +session_id = rag_agent.create_session(session_name=f"s{uuid.uuid4().hex}") + +turns = ["what is torchtune", "tell me about dora"] + +for t in turns: + print("user>", t) + stream = rag_agent.create_turn( + messages=[{"role": "user", "content": t}], session_id=session_id, stream=True + ) + for event in AgentEventLogger().log(stream): + event.print() +``` +#### ii. Run the Script +Let's run the script using `uv` +```bash +uv run python rag_agent.py +``` + +```{dropdown} ๐Ÿ‘‹ Click here to see the sample output + user> what is torchtune + inference> [knowledge_search(query='TorchTune')] + tool_execution> Tool:knowledge_search Args:{'query': 'TorchTune'} + tool_execution> Tool:knowledge_search Response:[TextContentItem(text='knowledge_search tool found 5 chunks:\nBEGIN of knowledge_search tool results.\n', type='text'), TextContentItem(text='Result 1:\nDocument_id:num-1\nContent: conversational data, :func:`~torchtune.datasets.chat_dataset` seems to be a good fit. ..., type='text'), TextContentItem(text='END of knowledge_search tool results.\n', type='text')] + inference> Here is a high-level overview of the text: + + **LoRA Finetuning with PyTorch Tune** + + PyTorch Tune provides a recipe for LoRA (Low-Rank Adaptation) finetuning, which is a technique to adapt pre-trained models to new tasks. The recipe uses the `lora_finetune_distributed` command. + ... + Overall, DORA is a powerful reinforcement learning algorithm that can learn complex tasks from human demonstrations. However, it requires careful consideration of the challenges and limitations to achieve optimal results. +``` + + + +**You're Ready to Build Your Own Apps!** + +Congrats! ๐Ÿฅณ Now you're ready to [build your own Llama Stack applications](../building_applications/)! ๐Ÿš€ diff --git a/versioned_docs/version-v0.2.23/getting_started/libraries.mdx b/versioned_docs/version-v0.2.23/getting_started/libraries.mdx new file mode 100644 index 0000000..7cbb792 --- /dev/null +++ b/versioned_docs/version-v0.2.23/getting_started/libraries.mdx @@ -0,0 +1,16 @@ +--- +description: We have a number of client-side SDKs available for different languages. +sidebar_label: Libraries +sidebar_position: 2 +title: Libraries (SDKs) +--- +## Libraries (SDKs) + +We have a number of client-side SDKs available for different languages. + +| **Language** | **Client SDK** | **Package** | +| :----: | :----: | :----: | +| Python | [llama-stack-client-python](https://github.com/meta-llama/llama-stack-client-python) | [![PyPI version](https://img.shields.io/pypi/v/llama_stack_client.svg)](https://pypi.org/project/llama_stack_client/) +| Swift | [llama-stack-client-swift](https://github.com/meta-llama/llama-stack-client-swift/tree/latest-release) | [![Swift Package Index](https://img.shields.io/endpoint?url=https%3A%2F%2Fswiftpackageindex.com%2Fapi%2Fpackages%2Fmeta-llama%2Fllama-stack-client-swift%2Fbadge%3Ftype%3Dswift-versions)](https://swiftpackageindex.com/meta-llama/llama-stack-client-swift) +| Node | [llama-stack-client-node](https://github.com/meta-llama/llama-stack-client-node) | [![NPM version](https://img.shields.io/npm/v/llama-stack-client.svg)](https://npmjs.org/package/llama-stack-client) +| Kotlin | [llama-stack-client-kotlin](https://github.com/meta-llama/llama-stack-client-kotlin/tree/latest-release) | [![Maven version](https://img.shields.io/maven-central/v/com.llama.llamastack/llama-stack-client-kotlin)](https://central.sonatype.com/artifact/com.llama.llamastack/llama-stack-client-kotlin) diff --git a/versioned_docs/version-v0.2.23/getting_started/quickstart.mdx b/versioned_docs/version-v0.2.23/getting_started/quickstart.mdx new file mode 100644 index 0000000..b885f3c --- /dev/null +++ b/versioned_docs/version-v0.2.23/getting_started/quickstart.mdx @@ -0,0 +1,149 @@ +--- +description: environments. +sidebar_label: Quickstart +sidebar_position: 1 +title: Quickstart +--- + +Get started with Llama Stack in minutes! + +Llama Stack is a stateful service with REST APIs to support the seamless transition of AI applications across different +environments. You can build and test using a local server first and deploy to a hosted endpoint for production. + +In this guide, we'll walk through how to build a RAG application locally using Llama Stack with [Ollama](https://ollama.com/) +as the inference [provider](/docs/providers/inference) for a Llama Model. + +**๐Ÿ’ก Notebook Version:** You can also follow this quickstart guide in a Jupyter notebook format: [quick_start.ipynb](https://github.com/meta-llama/llama-stack/blob/main/docs/quick_start.ipynb) + +#### Step 1: Install and setup +1. Install [uv](https://docs.astral.sh/uv/) +2. Run inference on a Llama model with [Ollama](https://ollama.com/download) +```bash +ollama run llama3.2:3b --keepalive 60m +``` + +#### Step 2: Run the Llama Stack server + +We will use `uv` to run the Llama Stack server. +```bash +OLLAMA_URL=http://localhost:11434 \ + uv run --with llama-stack llama stack build --distro starter --image-type venv --run +``` +#### Step 3: Run the demo +Now open up a new terminal and copy the following script into a file named `demo_script.py`. + +```python title="demo_script.py" +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +from llama_stack_client import Agent, AgentEventLogger, RAGDocument, LlamaStackClient + +vector_db_id = "my_demo_vector_db" +client = LlamaStackClient(base_url="http://localhost:8321") + +models = client.models.list() + +# Select the first LLM and first embedding models +model_id = next(m for m in models if m.model_type == "llm").identifier +embedding_model_id = ( + em := next(m for m in models if m.model_type == "embedding") +).identifier +embedding_dimension = em.metadata["embedding_dimension"] + +vector_db = client.vector_dbs.register( + vector_db_id=vector_db_id, + embedding_model=embedding_model_id, + embedding_dimension=embedding_dimension, + provider_id="faiss", +) +vector_db_id = vector_db.identifier +source = "https://www.paulgraham.com/greatwork.html" +print("rag_tool> Ingesting document:", source) +document = RAGDocument( + document_id="document_1", + content=source, + mime_type="text/html", + metadata={}, +) +client.tool_runtime.rag_tool.insert( + documents=[document], + vector_db_id=vector_db_id, + chunk_size_in_tokens=100, +) +agent = Agent( + client, + model=model_id, + instructions="You are a helpful assistant", + tools=[ + { + "name": "builtin::rag/knowledge_search", + "args": {"vector_db_ids": [vector_db_id]}, + } + ], +) + +prompt = "How do you do great work?" +print("prompt>", prompt) + +use_stream = True +response = agent.create_turn( + messages=[{"role": "user", "content": prompt}], + session_id=agent.create_session("rag_session"), + stream=use_stream, +) + +# Only call `AgentEventLogger().log(response)` for streaming responses. +if use_stream: + for log in AgentEventLogger().log(response): + log.print() +else: + print(response) +``` +We will use `uv` to run the script +``` +uv run --with llama-stack-client,fire,requests demo_script.py +``` +And you should see output like below. +``` +rag_tool> Ingesting document: https://www.paulgraham.com/greatwork.html + +prompt> How do you do great work? + +inference> [knowledge_search(query="What is the key to doing great work")] + +tool_execution> Tool:knowledge_search Args:{'query': 'What is the key to doing great work'} + +tool_execution> Tool:knowledge_search Response:[TextContentItem(text='knowledge_search tool found 5 chunks:\nBEGIN of knowledge_search tool results.\n', type='text'), TextContentItem(text="Result 1:\nDocument_id:docum\nContent: work. Doing great work means doing something important\nso well that you expand people's ideas of what's possible. But\nthere's no threshold for importance. It's a matter of degree, and\noften hard to judge at the time anyway.\n", type='text'), TextContentItem(text="Result 2:\nDocument_id:docum\nContent: work. Doing great work means doing something important\nso well that you expand people's ideas of what's possible. But\nthere's no threshold for importance. It's a matter of degree, and\noften hard to judge at the time anyway.\n", type='text'), TextContentItem(text="Result 3:\nDocument_id:docum\nContent: work. Doing great work means doing something important\nso well that you expand people's ideas of what's possible. But\nthere's no threshold for importance. It's a matter of degree, and\noften hard to judge at the time anyway.\n", type='text'), TextContentItem(text="Result 4:\nDocument_id:docum\nContent: work. Doing great work means doing something important\nso well that you expand people's ideas of what's possible. But\nthere's no threshold for importance. It's a matter of degree, and\noften hard to judge at the time anyway.\n", type='text'), TextContentItem(text="Result 5:\nDocument_id:docum\nContent: work. Doing great work means doing something important\nso well that you expand people's ideas of what's possible. But\nthere's no threshold for importance. It's a matter of degree, and\noften hard to judge at the time anyway.\n", type='text'), TextContentItem(text='END of knowledge_search tool results.\n', type='text')] + +inference> Based on the search results, it seems that doing great work means doing something important so well that you expand people's ideas of what's possible. However, there is no clear threshold for importance, and it can be difficult to judge at the time. + +To further clarify, I would suggest that doing great work involves: + +* Completing tasks with high quality and attention to detail +* Expanding on existing knowledge or ideas +* Making a positive impact on others through your work +* Striving for excellence and continuous improvement + +Ultimately, great work is about making a meaningful contribution and leaving a lasting impression. +``` +Congratulations! You've successfully built your first RAG application using Llama Stack! ๐ŸŽ‰๐Ÿฅณ + +:::tip HuggingFace access + +If you are getting a **401 Client Error** from HuggingFace for the **all-MiniLM-L6-v2** model, try setting **HF_TOKEN** to a valid HuggingFace token in your environment + +::: + +### Next Steps + +Now you're ready to dive deeper into Llama Stack! +- Explore the [Detailed Tutorial](./detailed_tutorial). +- Try the [Getting Started Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb). +- Browse more [Notebooks on GitHub](https://github.com/meta-llama/llama-stack/tree/main/docs/notebooks). +- Learn about Llama Stack [Concepts](/docs/concepts). +- Discover how to [Build Llama Stacks](/docs/distributions). +- Refer to our [References](/docs/references) for details on the Llama CLI and Python SDK. +- Check out the [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repository for example applications and tutorials. diff --git a/versioned_docs/version-v0.2.23/index.mdx b/versioned_docs/version-v0.2.23/index.mdx new file mode 100644 index 0000000..bed931f --- /dev/null +++ b/versioned_docs/version-v0.2.23/index.mdx @@ -0,0 +1,101 @@ +--- +sidebar_position: 1 +title: Welcome to Llama Stack +description: Llama Stack is the open-source framework for building generative AI applications +sidebar_label: Intro +tags: + - getting-started + - overview +--- + +# Welcome to Llama Stack + +Llama Stack is the open-source framework for building generative AI applications. + +:::tip Llama 4 is here! + +Check out [Getting Started with Llama 4](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started_llama4.ipynb) + +::: + +:::tip News + +Llama Stack is now available! See the [release notes](https://github.com/meta-llama/llama-stack/releases) for more details. + +::: + + +## What is Llama Stack? + +Llama Stack defines and standardizes the core building blocks needed to bring generative AI applications to market. It provides a unified set of APIs with implementations from leading service providers, enabling seamless transitions between development and production environments. More specifically, it provides: + +- **Unified API layer** for Inference, RAG, Agents, Tools, Safety, Evals, and Telemetry. +- **Plugin architecture** to support the rich ecosystem of implementations of the different APIs in different environments like local development, on-premises, cloud, and mobile. +- **Prepackaged verified distributions** which offer a one-stop solution for developers to get started quickly and reliably in any environment +- **Multiple developer interfaces** like CLI and SDKs for Python, Node, iOS, and Android +- **Standalone applications** as examples for how to build production-grade AI applications with Llama Stack + +Llama Stack + +Our goal is to provide pre-packaged implementations (aka "distributions") which can be run in a variety of deployment environments. LlamaStack can assist you in your entire app development lifecycle - start iterating on local, mobile or desktop and seamlessly transition to on-prem or public cloud deployments. At every point in this transition, the same set of APIs and the same developer experience is available. + +## How does Llama Stack work? + +Llama Stack consists of a server (with multiple pluggable API providers) and Client SDKs meant to be used in your applications. The server can be run in a variety of environments, including local (inline) development, on-premises, and cloud. The client SDKs are available for Python, Swift, Node, and Kotlin. + +## Quick Links + +- Ready to build? Check out the [Getting Started Guide](https://llama-stack.github.io/getting_started/quickstart) to get started. +- Want to contribute? See the [Contributing Guide](https://github.com/llamastack/llama-stack/blob/main/CONTRIBUTING.md). +- Explore [Example Applications](https://github.com/llamastack/llama-stack-apps) built with Llama Stack. + +## Rich Ecosystem Support + +Llama Stack provides adapters for popular providers across all API categories: + +- **Inference**: Meta Reference, Ollama, Fireworks, Together, NVIDIA, vLLM, AWS Bedrock, OpenAI, Anthropic, and more +- **Vector Databases**: FAISS, Chroma, Milvus, Postgres, Weaviate, Qdrant, and others +- **Safety**: Llama Guard, Prompt Guard, Code Scanner, AWS Bedrock +- **Training & Evaluation**: HuggingFace, TorchTune, NVIDIA NEMO + +:::info Provider Details +For complete provider compatibility and setup instructions, see our [Providers Documentation](https://llamastack.github.io/providers/). +::: + +## Get Started Today + + diff --git a/versioned_docs/version-v0.2.23/providers/agents/index.mdx b/versioned_docs/version-v0.2.23/providers/agents/index.mdx new file mode 100644 index 0000000..5cd3777 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/agents/index.mdx @@ -0,0 +1,27 @@ +--- +description: "Agents API for creating and interacting with agentic systems. + + Main functionalities provided by this API: + - Create agents with specific instructions and ability to use tools. + - Interactions with agents are grouped into sessions (\"threads\"), and each interaction is called a \"turn\". + - Agents can be provided with various tools (see the ToolGroups and ToolRuntime APIs for more details). + - Agents can be provided with various shields (see the Safety API for more details). + - Agents can also use Memory to retrieve information from knowledge bases. See the RAG Tool and Vector IO APIs for more details." +sidebar_label: Agents +title: Agents +--- + +# Agents + +## Overview + +Agents API for creating and interacting with agentic systems. + + Main functionalities provided by this API: + - Create agents with specific instructions and ability to use tools. + - Interactions with agents are grouped into sessions ("threads"), and each interaction is called a "turn". + - Agents can be provided with various tools (see the ToolGroups and ToolRuntime APIs for more details). + - Agents can be provided with various shields (see the Safety API for more details). + - Agents can also use Memory to retrieve information from knowledge bases. See the RAG Tool and Vector IO APIs for more details. + +This section contains documentation for all available providers for the **agents** API. diff --git a/versioned_docs/version-v0.2.23/providers/agents/inline_meta-reference.mdx b/versioned_docs/version-v0.2.23/providers/agents/inline_meta-reference.mdx new file mode 100644 index 0000000..fd96174 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/agents/inline_meta-reference.mdx @@ -0,0 +1,29 @@ +--- +description: "Meta's reference implementation of an agent system that can use tools, access vector databases, and perform complex reasoning tasks." +sidebar_label: Meta-Reference +title: inline::meta-reference +--- + +# inline::meta-reference + +## Description + +Meta's reference implementation of an agent system that can use tools, access vector databases, and perform complex reasoning tasks. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `persistence_store` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | +| `responses_store` | `utils.sqlstore.sqlstore.SqliteSqlStoreConfig \| utils.sqlstore.sqlstore.PostgresSqlStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +persistence_store: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/agents_store.db +responses_store: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/responses_store.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/batches/index.mdx b/versioned_docs/version-v0.2.23/providers/batches/index.mdx new file mode 100644 index 0000000..2c64b27 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/batches/index.mdx @@ -0,0 +1,31 @@ +--- +description: "The Batches API enables efficient processing of multiple requests in a single operation, + particularly useful for processing large datasets, batch evaluation workflows, and + cost-effective inference at scale. + + The API is designed to allow use of openai client libraries for seamless integration. + + This API provides the following extensions: + - idempotent batch creation + + Note: This API is currently under active development and may undergo changes." +sidebar_label: Batches +title: Batches +--- + +# Batches + +## Overview + +The Batches API enables efficient processing of multiple requests in a single operation, + particularly useful for processing large datasets, batch evaluation workflows, and + cost-effective inference at scale. + + The API is designed to allow use of openai client libraries for seamless integration. + + This API provides the following extensions: + - idempotent batch creation + + Note: This API is currently under active development and may undergo changes. + +This section contains documentation for all available providers for the **batches** API. diff --git a/versioned_docs/version-v0.2.23/providers/batches/inline_reference.mdx b/versioned_docs/version-v0.2.23/providers/batches/inline_reference.mdx new file mode 100644 index 0000000..f438005 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/batches/inline_reference.mdx @@ -0,0 +1,27 @@ +--- +description: "Reference implementation of batches API with KVStore persistence." +sidebar_label: Reference +title: inline::reference +--- + +# inline::reference + +## Description + +Reference implementation of batches API with KVStore persistence. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Configuration for the key-value store backend. | +| `max_concurrent_batches` | `` | No | 1 | Maximum number of concurrent batches to process simultaneously. | +| `max_concurrent_requests_per_batch` | `` | No | 10 | Maximum number of concurrent requests to process per batch. | + +## Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/batches.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/datasetio/index.mdx b/versioned_docs/version-v0.2.23/providers/datasetio/index.mdx new file mode 100644 index 0000000..aeeb019 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/datasetio/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Datasetio +title: Datasetio +--- + +# Datasetio + +## Overview + +This section contains documentation for all available providers for the **datasetio** API. diff --git a/versioned_docs/version-v0.2.23/providers/datasetio/inline_localfs.mdx b/versioned_docs/version-v0.2.23/providers/datasetio/inline_localfs.mdx new file mode 100644 index 0000000..b02a3a3 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/datasetio/inline_localfs.mdx @@ -0,0 +1,25 @@ +--- +description: "Local filesystem-based dataset I/O provider for reading and writing datasets to local storage." +sidebar_label: Localfs +title: inline::localfs +--- + +# inline::localfs + +## Description + +Local filesystem-based dataset I/O provider for reading and writing datasets to local storage. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/localfs_datasetio.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/datasetio/remote_huggingface.mdx b/versioned_docs/version-v0.2.23/providers/datasetio/remote_huggingface.mdx new file mode 100644 index 0000000..82597d9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/datasetio/remote_huggingface.mdx @@ -0,0 +1,25 @@ +--- +description: "HuggingFace datasets provider for accessing and managing datasets from the HuggingFace Hub." +sidebar_label: Remote - Huggingface +title: remote::huggingface +--- + +# remote::huggingface + +## Description + +HuggingFace datasets provider for accessing and managing datasets from the HuggingFace Hub. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/huggingface_datasetio.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/datasetio/remote_nvidia.mdx b/versioned_docs/version-v0.2.23/providers/datasetio/remote_nvidia.mdx new file mode 100644 index 0000000..35a7dac --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/datasetio/remote_nvidia.mdx @@ -0,0 +1,29 @@ +--- +description: "NVIDIA's dataset I/O provider for accessing datasets from NVIDIA's data platform." +sidebar_label: Remote - Nvidia +title: remote::nvidia +--- + +# remote::nvidia + +## Description + +NVIDIA's dataset I/O provider for accessing datasets from NVIDIA's data platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The NVIDIA API key. | +| `dataset_namespace` | `str \| None` | No | default | The NVIDIA dataset namespace. | +| `project_id` | `str \| None` | No | test-project | The NVIDIA project ID. | +| `datasets_url` | `` | No | http://nemo.test | Base URL for the NeMo Dataset API | + +## Sample Configuration + +```yaml +api_key: ${env.NVIDIA_API_KEY:=} +dataset_namespace: ${env.NVIDIA_DATASET_NAMESPACE:=default} +project_id: ${env.NVIDIA_PROJECT_ID:=test-project} +datasets_url: ${env.NVIDIA_DATASETS_URL:=http://nemo.test} +``` diff --git a/versioned_docs/version-v0.2.23/providers/eval/index.mdx b/versioned_docs/version-v0.2.23/providers/eval/index.mdx new file mode 100644 index 0000000..73b0b89 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/eval/index.mdx @@ -0,0 +1,13 @@ +--- +description: "Llama Stack Evaluation API for running evaluations on model and agent candidates." +sidebar_label: Eval +title: Eval +--- + +# Eval + +## Overview + +Llama Stack Evaluation API for running evaluations on model and agent candidates. + +This section contains documentation for all available providers for the **eval** API. diff --git a/versioned_docs/version-v0.2.23/providers/eval/inline_meta-reference.mdx b/versioned_docs/version-v0.2.23/providers/eval/inline_meta-reference.mdx new file mode 100644 index 0000000..b0eb589 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/eval/inline_meta-reference.mdx @@ -0,0 +1,25 @@ +--- +description: "Meta's reference implementation of evaluation tasks with support for multiple languages and evaluation metrics." +sidebar_label: Meta-Reference +title: inline::meta-reference +--- + +# inline::meta-reference + +## Description + +Meta's reference implementation of evaluation tasks with support for multiple languages and evaluation metrics. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/meta_reference_eval.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/eval/remote_nvidia.mdx b/versioned_docs/version-v0.2.23/providers/eval/remote_nvidia.mdx new file mode 100644 index 0000000..36bb472 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/eval/remote_nvidia.mdx @@ -0,0 +1,23 @@ +--- +description: "NVIDIA's evaluation provider for running evaluation tasks on NVIDIA's platform." +sidebar_label: Remote - Nvidia +title: remote::nvidia +--- + +# remote::nvidia + +## Description + +NVIDIA's evaluation provider for running evaluation tasks on NVIDIA's platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `evaluator_url` | `` | No | http://0.0.0.0:7331 | The url for accessing the evaluator service | + +## Sample Configuration + +```yaml +evaluator_url: ${env.NVIDIA_EVALUATOR_URL:=http://localhost:7331} +``` diff --git a/versioned_docs/version-v0.2.23/providers/external/external-providers-guide.mdx b/versioned_docs/version-v0.2.23/providers/external/external-providers-guide.mdx new file mode 100644 index 0000000..eb30afd --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/external/external-providers-guide.mdx @@ -0,0 +1,286 @@ +# Creating External Providers + +## Configuration + +To enable external providers, you need to add `module` into your build yaml, allowing Llama Stack to install the required package corresponding to the external provider. + +an example entry in your build.yaml should look like: + +``` +- provider_type: remote::ramalama + module: ramalama_stack +``` + +Additionally you can configure the `external_providers_dir` in your Llama Stack configuration. This method is in the process of being deprecated in favor of the `module` method. If using this method, the external provider directory should contain your external provider specifications: + +```yaml +external_providers_dir: ~/.llama/providers.d/ +``` + +## Directory Structure + +The external providers directory should follow this structure: + +``` +providers.d/ + remote/ + inference/ + custom_ollama.yaml + vllm.yaml + vector_io/ + qdrant.yaml + safety/ + llama-guard.yaml + inline/ + inference/ + custom_ollama.yaml + vllm.yaml + vector_io/ + qdrant.yaml + safety/ + llama-guard.yaml +``` + +Each YAML file in these directories defines a provider specification for that particular API. + +## Provider Types + +Llama Stack supports two types of external providers: + +1. **Remote Providers**: Providers that communicate with external services (e.g., cloud APIs) +2. **Inline Providers**: Providers that run locally within the Llama Stack process + +### Remote Provider Specification + +Remote providers are used when you need to communicate with external services. Here's an example for a custom Ollama provider: + +```yaml +adapter: + adapter_type: custom_ollama + pip_packages: + - ollama + - aiohttp + config_class: llama_stack_ollama_provider.config.OllamaImplConfig + module: llama_stack_ollama_provider +api_dependencies: [] +optional_api_dependencies: [] +``` + +#### Adapter Configuration + +The `adapter` section defines how to load and configure the provider: + +- `adapter_type`: A unique identifier for this adapter +- `pip_packages`: List of Python packages required by the provider +- `config_class`: The full path to the configuration class +- `module`: The Python module containing the provider implementation + +### Inline Provider Specification + +Inline providers run locally within the Llama Stack process. Here's an example for a custom vector store provider: + +```yaml +module: llama_stack_vector_provider +config_class: llama_stack_vector_provider.config.VectorStoreConfig +pip_packages: + - faiss-cpu + - numpy +api_dependencies: + - inference +optional_api_dependencies: + - vector_io +provider_data_validator: llama_stack_vector_provider.validator.VectorStoreValidator +container_image: custom-vector-store:latest # optional +``` + +#### Inline Provider Fields + +- `module`: The Python module containing the provider implementation +- `config_class`: The full path to the configuration class +- `pip_packages`: List of Python packages required by the provider +- `api_dependencies`: List of Llama Stack APIs that this provider depends on +- `optional_api_dependencies`: List of optional Llama Stack APIs that this provider can use +- `provider_data_validator`: Optional validator for provider data +- `container_image`: Optional container image to use instead of pip packages + +## Required Fields + +### All Providers + +All providers must contain a `get_provider_spec` function in their `provider` module. This is a standardized structure that Llama Stack expects and is necessary for getting things such as the config class. The `get_provider_spec` method returns a structure identical to the `adapter`. An example function may look like: + +```python +from llama_stack.providers.datatypes import ( + ProviderSpec, + Api, + AdapterSpec, + remote_provider_spec, +) + + +def get_provider_spec() -> ProviderSpec: + return remote_provider_spec( + api=Api.inference, + adapter=AdapterSpec( + adapter_type="ramalama", + pip_packages=["ramalama>=0.8.5", "pymilvus"], + config_class="ramalama_stack.config.RamalamaImplConfig", + module="ramalama_stack", + ), + ) +``` + +#### Remote Providers + +Remote providers must expose a `get_adapter_impl()` function in their module that takes two arguments: +1. `config`: An instance of the provider's config class +2. `deps`: A dictionary of API dependencies + +This function must return an instance of the provider's adapter class that implements the required protocol for the API. + +Example: +```python +async def get_adapter_impl( + config: OllamaImplConfig, deps: Dict[Api, Any] +) -> OllamaInferenceAdapter: + return OllamaInferenceAdapter(config) +``` + +#### Inline Providers + +Inline providers must expose a `get_provider_impl()` function in their module that takes two arguments: +1. `config`: An instance of the provider's config class +2. `deps`: A dictionary of API dependencies + +Example: +```python +async def get_provider_impl( + config: VectorStoreConfig, deps: Dict[Api, Any] +) -> VectorStoreImpl: + impl = VectorStoreImpl(config, deps[Api.inference]) + await impl.initialize() + return impl +``` + +## Dependencies + +The provider package must be installed on the system. For example: + +```bash +$ uv pip show llama-stack-ollama-provider +Name: llama-stack-ollama-provider +Version: 0.1.0 +Location: /path/to/venv/lib/python3.10/site-packages +``` + +## Best Practices + +1. **Package Naming**: Use the prefix `llama-stack-provider-` for your provider packages to make them easily identifiable. + +2. **Version Management**: Keep your provider package versioned and compatible with the Llama Stack version you're using. + +3. **Dependencies**: Only include the minimum required dependencies in your provider package. + +4. **Documentation**: Include clear documentation in your provider package about: + - Installation requirements + - Configuration options + - Usage examples + - Any limitations or known issues + +5. **Testing**: Include tests in your provider package to ensure it works correctly with Llama Stack. +You can refer to the [integration tests +guide](https://github.com/meta-llama/llama-stack/blob/main/tests/integration/README.md) for more +information. Execute the test for the Provider type you are developing. + +## Troubleshooting + +If your external provider isn't being loaded: + +1. Check that `module` points to a published pip package with a top level `provider` module including `get_provider_spec`. +1. Check that the `external_providers_dir` path is correct and accessible. +2. Verify that the YAML files are properly formatted. +3. Ensure all required Python packages are installed. +4. Check the Llama Stack server logs for any error messages - turn on debug logging to get more + information using `LLAMA_STACK_LOGGING=all=debug`. +5. Verify that the provider package is installed in your Python environment if using `external_providers_dir`. + +## Examples + +### Example using `external_providers_dir`: Custom Ollama Provider + +Here's a complete example of creating and using a custom Ollama provider: + +1. First, create the provider package: + +```bash +mkdir -p llama-stack-provider-ollama +cd llama-stack-provider-ollama +git init +uv init +``` + +2. Edit `pyproject.toml`: + +```toml +[project] +name = "llama-stack-provider-ollama" +version = "0.1.0" +description = "Ollama provider for Llama Stack" +requires-python = ">=3.12" +dependencies = ["llama-stack", "pydantic", "ollama", "aiohttp"] +``` + +3. Create the provider specification: + +```yaml +# ~/.llama/providers.d/remote/inference/custom_ollama.yaml +adapter: + adapter_type: custom_ollama + pip_packages: ["ollama", "aiohttp"] + config_class: llama_stack_provider_ollama.config.OllamaImplConfig + module: llama_stack_provider_ollama +api_dependencies: [] +optional_api_dependencies: [] +``` + +4. Install the provider: + +```bash +uv pip install -e . +``` + +5. Configure Llama Stack to use external providers: + +```yaml +external_providers_dir: ~/.llama/providers.d/ +``` + +The provider will now be available in Llama Stack with the type `remote::custom_ollama`. + + +### Example using `module`: ramalama-stack + +[ramalama-stack](https://github.com/containers/ramalama-stack) is a recognized external provider that supports installation via module. + +To install Llama Stack with this external provider a user can provider the following build.yaml: + +```yaml +version: 2 +distribution_spec: + description: Use (an external) Ramalama server for running LLM inference + container_image: null + providers: + inference: + - provider_type: remote::ramalama + module: ramalama_stack==0.3.0a0 +image_type: venv +image_name: null +external_providers_dir: null +additional_pip_packages: +- aiosqlite +- sqlalchemy[asyncio] +``` + +No other steps are required other than `llama stack build` and `llama stack run`. The build process will use `module` to install all of the provider dependencies, retrieve the spec, etc. + +The provider will now be available in Llama Stack with the type `remote::ramalama`. diff --git a/versioned_docs/version-v0.2.23/providers/external/external-providers-list.mdx b/versioned_docs/version-v0.2.23/providers/external/external-providers-list.mdx new file mode 100644 index 0000000..45fcc50 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/external/external-providers-list.mdx @@ -0,0 +1,11 @@ +# Known External Providers + +Here's a list of known external providers that you can use with Llama Stack: + +| Name | Description | API | Type | Repository | +|------|-------------|-----|------|------------| +| KubeFlow Training | Train models with KubeFlow | Post Training | Remote | [llama-stack-provider-kft](https://github.com/opendatahub-io/llama-stack-provider-kft) | +| KubeFlow Pipelines | Train models with KubeFlow Pipelines | Post Training | Inline **and** Remote | [llama-stack-provider-kfp-trainer](https://github.com/opendatahub-io/llama-stack-provider-kfp-trainer) | +| RamaLama | Inference models with RamaLama | Inference | Remote | [ramalama-stack](https://github.com/containers/ramalama-stack) | +| TrustyAI LM-Eval | Evaluate models with TrustyAI LM-Eval | Eval | Remote | [llama-stack-provider-lmeval](https://github.com/trustyai-explainability/llama-stack-provider-lmeval) | +| MongoDB | VectorIO with MongoDB | Vector_IO | Remote | [mongodb-llama-stack](https://github.com/mongodb-partners/mongodb-llama-stack) | diff --git a/versioned_docs/version-v0.2.23/providers/external/index.mdx b/versioned_docs/version-v0.2.23/providers/external/index.mdx new file mode 100644 index 0000000..28a9a11 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/external/index.mdx @@ -0,0 +1,11 @@ +# External Providers + +Llama Stack supports external providers that live outside of the main codebase. This allows you to: +- Create and maintain your own providers independently +- Share providers with others without contributing to the main codebase +- Keep provider-specific code separate from the core Llama Stack code + +## External Provider Documentation + +- [Known External Providers](./external-providers-list.mdx) +- [Creating External Providers](./external-providers-guide.mdx) diff --git a/versioned_docs/version-v0.2.23/providers/files/index.mdx b/versioned_docs/version-v0.2.23/providers/files/index.mdx new file mode 100644 index 0000000..7d729d9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/files/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Files +title: Files +--- + +# Files + +## Overview + +This section contains documentation for all available providers for the **files** API. diff --git a/versioned_docs/version-v0.2.23/providers/files/inline_localfs.mdx b/versioned_docs/version-v0.2.23/providers/files/inline_localfs.mdx new file mode 100644 index 0000000..86d141f --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/files/inline_localfs.mdx @@ -0,0 +1,28 @@ +--- +description: "Local filesystem-based file storage provider for managing files and documents locally." +sidebar_label: Localfs +title: inline::localfs +--- + +# inline::localfs + +## Description + +Local filesystem-based file storage provider for managing files and documents locally. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `storage_dir` | `` | No | | Directory to store uploaded files | +| `metadata_store` | `utils.sqlstore.sqlstore.SqliteSqlStoreConfig \| utils.sqlstore.sqlstore.PostgresSqlStoreConfig` | No | sqlite | SQL store configuration for file metadata | +| `ttl_secs` | `` | No | 31536000 | | + +## Sample Configuration + +```yaml +storage_dir: ${env.FILES_STORAGE_DIR:=~/.llama/dummy/files} +metadata_store: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/files_metadata.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/files/remote_s3.mdx b/versioned_docs/version-v0.2.23/providers/files/remote_s3.mdx new file mode 100644 index 0000000..353cedb --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/files/remote_s3.mdx @@ -0,0 +1,37 @@ +--- +description: "AWS S3-based file storage provider for scalable cloud file management with metadata persistence." +sidebar_label: Remote - S3 +title: remote::s3 +--- + +# remote::s3 + +## Description + +AWS S3-based file storage provider for scalable cloud file management with metadata persistence. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `bucket_name` | `` | No | | S3 bucket name to store files | +| `region` | `` | No | us-east-1 | AWS region where the bucket is located | +| `aws_access_key_id` | `str \| None` | No | | AWS access key ID (optional if using IAM roles) | +| `aws_secret_access_key` | `str \| None` | No | | AWS secret access key (optional if using IAM roles) | +| `endpoint_url` | `str \| None` | No | | Custom S3 endpoint URL (for MinIO, LocalStack, etc.) | +| `auto_create_bucket` | `` | No | False | Automatically create the S3 bucket if it doesn't exist | +| `metadata_store` | `utils.sqlstore.sqlstore.SqliteSqlStoreConfig \| utils.sqlstore.sqlstore.PostgresSqlStoreConfig` | No | sqlite | SQL store configuration for file metadata | + +## Sample Configuration + +```yaml +bucket_name: ${env.S3_BUCKET_NAME} +region: ${env.AWS_REGION:=us-east-1} +aws_access_key_id: ${env.AWS_ACCESS_KEY_ID:=} +aws_secret_access_key: ${env.AWS_SECRET_ACCESS_KEY:=} +endpoint_url: ${env.S3_ENDPOINT_URL:=} +auto_create_bucket: ${env.S3_AUTO_CREATE_BUCKET:=false} +metadata_store: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/s3_files_metadata.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/index.mdx b/versioned_docs/version-v0.2.23/providers/index.mdx new file mode 100644 index 0000000..9c560fe --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/index.mdx @@ -0,0 +1,33 @@ +--- +title: API Providers +description: Ecosystem of providers for swapping implementations across the same API +sidebar_label: Overview +sidebar_position: 1 +--- + +# API Providers + +The goal of Llama Stack is to build an ecosystem where users can easily swap out different implementations for the same API. Examples for these include: +- LLM inference providers (e.g., Meta Reference, Ollama, Fireworks, Together, AWS Bedrock, Groq, Cerebras, SambaNova, vLLM, OpenAI, Anthropic, Gemini, WatsonX, etc.), +- Vector databases (e.g., FAISS, SQLite-Vec, ChromaDB, Weaviate, Qdrant, Milvus, PGVector, etc.), +- Safety providers (e.g., Meta's Llama Guard, Prompt Guard, Code Scanner, AWS Bedrock Guardrails, etc.), +- Tool Runtime providers (e.g., RAG Runtime, Brave Search, etc.) + +Providers come in two flavors: +- **Remote**: the provider runs as a separate service external to the Llama Stack codebase. Llama Stack contains a small amount of adapter code. +- **Inline**: the provider is fully specified and implemented within the Llama Stack codebase. It may be a simple wrapper around an existing library, or a full fledged implementation within Llama Stack. + +Importantly, Llama Stack always strives to provide at least one fully inline provider for each API so you can iterate on a fully featured environment locally. + +## Provider Categories + +- **[External Providers](external/index.mdx)** - Guide for building and using external providers +- **[OpenAI Compatibility](./openai.mdx)** - OpenAI API compatibility layer +- **[Inference](inference/index.mdx)** - LLM and embedding model providers +- **[Agents](agents/index.mdx)** - Agentic system providers +- **[DatasetIO](datasetio/index.mdx)** - Dataset and data loader providers +- **[Safety](safety/index.mdx)** - Content moderation and safety providers +- **[Telemetry](telemetry/index.mdx)** - Monitoring and observability providers +- **[Vector IO](vector_io/index.mdx)** - Vector database providers +- **[Tool Runtime](tool_runtime/index.mdx)** - Tool and protocol providers +- **[Files](files/index.mdx)** - File system and storage providers diff --git a/versioned_docs/version-v0.2.23/providers/inference/index.mdx b/versioned_docs/version-v0.2.23/providers/inference/index.mdx new file mode 100644 index 0000000..ebbaf1b --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/index.mdx @@ -0,0 +1,21 @@ +--- +description: "Llama Stack Inference API for generating completions, chat completions, and embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: + - LLM models: these models generate \"raw\" and \"chat\" (conversational) completions. + - Embedding models: these models generate embeddings to be used for semantic search." +sidebar_label: Inference +title: Inference +--- + +# Inference + +## Overview + +Llama Stack Inference API for generating completions, chat completions, and embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: + - LLM models: these models generate "raw" and "chat" (conversational) completions. + - Embedding models: these models generate embeddings to be used for semantic search. + +This section contains documentation for all available providers for the **inference** API. diff --git a/versioned_docs/version-v0.2.23/providers/inference/inline_meta-reference.mdx b/versioned_docs/version-v0.2.23/providers/inference/inline_meta-reference.mdx new file mode 100644 index 0000000..328586f --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/inline_meta-reference.mdx @@ -0,0 +1,36 @@ +--- +description: "Meta's reference implementation of inference with support for various model formats and optimization techniques." +sidebar_label: Meta-Reference +title: inline::meta-reference +--- + +# inline::meta-reference + +## Description + +Meta's reference implementation of inference with support for various model formats and optimization techniques. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `model` | `str \| None` | No | | | +| `torch_seed` | `int \| None` | No | | | +| `max_seq_len` | `` | No | 4096 | | +| `max_batch_size` | `` | No | 1 | | +| `model_parallel_size` | `int \| None` | No | | | +| `create_distributed_process_group` | `` | No | True | | +| `checkpoint_dir` | `str \| None` | No | | | +| `quantization` | `Bf16QuantizationConfig \| Fp8QuantizationConfig \| Int4QuantizationConfig, annotation=NoneType, required=True, discriminator='type'` | No | | | + +## Sample Configuration + +```yaml +model: Llama3.2-3B-Instruct +checkpoint_dir: ${env.CHECKPOINT_DIR:=null} +quantization: + type: ${env.QUANTIZATION_TYPE:=bf16} +model_parallel_size: ${env.MODEL_PARALLEL_SIZE:=0} +max_batch_size: ${env.MAX_BATCH_SIZE:=1} +max_seq_len: ${env.MAX_SEQ_LEN:=4096} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/inline_sentence-transformers.mdx b/versioned_docs/version-v0.2.23/providers/inference/inline_sentence-transformers.mdx new file mode 100644 index 0000000..0e207bb --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/inline_sentence-transformers.mdx @@ -0,0 +1,17 @@ +--- +description: "Sentence Transformers inference provider for text embeddings and similarity search." +sidebar_label: Sentence-Transformers +title: inline::sentence-transformers +--- + +# inline::sentence-transformers + +## Description + +Sentence Transformers inference provider for text embeddings and similarity search. + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_anthropic.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_anthropic.mdx new file mode 100644 index 0000000..6bd636c --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_anthropic.mdx @@ -0,0 +1,23 @@ +--- +description: "Anthropic inference provider for accessing Claude models and Anthropic's AI services." +sidebar_label: Remote - Anthropic +title: remote::anthropic +--- + +# remote::anthropic + +## Description + +Anthropic inference provider for accessing Claude models and Anthropic's AI services. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | API key for Anthropic models | + +## Sample Configuration + +```yaml +api_key: ${env.ANTHROPIC_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_azure.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_azure.mdx new file mode 100644 index 0000000..0eb0ea7 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_azure.mdx @@ -0,0 +1,36 @@ +--- +description: | + Azure OpenAI inference provider for accessing GPT models and other Azure services. + Provider documentation + https://learn.microsoft.com/en-us/azure/ai-foundry/openai/overview +sidebar_label: Remote - Azure +title: remote::azure +--- + +# remote::azure + +## Description + + +Azure OpenAI inference provider for accessing GPT models and other Azure services. +Provider documentation +https://learn.microsoft.com/en-us/azure/ai-foundry/openai/overview + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `` | No | | Azure API key for Azure | +| `api_base` | `` | No | | Azure API base for Azure (e.g., https://your-resource-name.openai.azure.com) | +| `api_version` | `str \| None` | No | | Azure API version for Azure (e.g., 2024-12-01-preview) | +| `api_type` | `str \| None` | No | azure | Azure API type for Azure (e.g., azure) | + +## Sample Configuration + +```yaml +api_key: ${env.AZURE_API_KEY:=} +api_base: ${env.AZURE_API_BASE:=} +api_version: ${env.AZURE_API_VERSION:=} +api_type: ${env.AZURE_API_TYPE:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_bedrock.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_bedrock.mdx new file mode 100644 index 0000000..04c2154 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_bedrock.mdx @@ -0,0 +1,32 @@ +--- +description: "AWS Bedrock inference provider for accessing various AI models through AWS's managed service." +sidebar_label: Remote - Bedrock +title: remote::bedrock +--- + +# remote::bedrock + +## Description + +AWS Bedrock inference provider for accessing various AI models through AWS's managed service. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `aws_access_key_id` | `str \| None` | No | | The AWS access key to use. Default use environment variable: AWS_ACCESS_KEY_ID | +| `aws_secret_access_key` | `str \| None` | No | | The AWS secret access key to use. Default use environment variable: AWS_SECRET_ACCESS_KEY | +| `aws_session_token` | `str \| None` | No | | The AWS session token to use. Default use environment variable: AWS_SESSION_TOKEN | +| `region_name` | `str \| None` | No | | The default AWS Region to use, for example, us-west-1 or us-west-2.Default use environment variable: AWS_DEFAULT_REGION | +| `profile_name` | `str \| None` | No | | The profile name that contains credentials to use.Default use environment variable: AWS_PROFILE | +| `total_max_attempts` | `int \| None` | No | | An integer representing the maximum number of attempts that will be made for a single request, including the initial attempt. Default use environment variable: AWS_MAX_ATTEMPTS | +| `retry_mode` | `str \| None` | No | | A string representing the type of retries Boto3 will perform.Default use environment variable: AWS_RETRY_MODE | +| `connect_timeout` | `float \| None` | No | 60.0 | The time in seconds till a timeout exception is thrown when attempting to make a connection. The default is 60 seconds. | +| `read_timeout` | `float \| None` | No | 60.0 | The time in seconds till a timeout exception is thrown when attempting to read from a connection.The default is 60 seconds. | +| `session_ttl` | `int \| None` | No | 3600 | The time in seconds till a session expires. The default is 3600 seconds (1 hour). | + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_cerebras.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_cerebras.mdx new file mode 100644 index 0000000..d9cc93a --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_cerebras.mdx @@ -0,0 +1,25 @@ +--- +description: "Cerebras inference provider for running models on Cerebras Cloud platform." +sidebar_label: Remote - Cerebras +title: remote::cerebras +--- + +# remote::cerebras + +## Description + +Cerebras inference provider for running models on Cerebras Cloud platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `base_url` | `` | No | https://api.cerebras.ai | Base URL for the Cerebras API | +| `api_key` | `` | No | | Cerebras API Key | + +## Sample Configuration + +```yaml +base_url: https://api.cerebras.ai +api_key: ${env.CEREBRAS_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_databricks.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_databricks.mdx new file mode 100644 index 0000000..7f736db --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_databricks.mdx @@ -0,0 +1,25 @@ +--- +description: "Databricks inference provider for running models on Databricks' unified analytics platform." +sidebar_label: Remote - Databricks +title: remote::databricks +--- + +# remote::databricks + +## Description + +Databricks inference provider for running models on Databricks' unified analytics platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | | The URL for the Databricks model serving endpoint | +| `api_token` | `` | No | | The Databricks API token | + +## Sample Configuration + +```yaml +url: ${env.DATABRICKS_HOST:=} +api_token: ${env.DATABRICKS_TOKEN:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_fireworks.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_fireworks.mdx new file mode 100644 index 0000000..d2c3a66 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_fireworks.mdx @@ -0,0 +1,26 @@ +--- +description: "Fireworks AI inference provider for Llama models and other AI models on the Fireworks platform." +sidebar_label: Remote - Fireworks +title: remote::fireworks +--- + +# remote::fireworks + +## Description + +Fireworks AI inference provider for Llama models and other AI models on the Fireworks platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. | +| `url` | `` | No | https://api.fireworks.ai/inference/v1 | The URL for the Fireworks server | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | The Fireworks.ai API Key | + +## Sample Configuration + +```yaml +url: https://api.fireworks.ai/inference/v1 +api_key: ${env.FIREWORKS_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_gemini.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_gemini.mdx new file mode 100644 index 0000000..0505c69 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_gemini.mdx @@ -0,0 +1,23 @@ +--- +description: "Google Gemini inference provider for accessing Gemini models and Google's AI services." +sidebar_label: Remote - Gemini +title: remote::gemini +--- + +# remote::gemini + +## Description + +Google Gemini inference provider for accessing Gemini models and Google's AI services. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | API key for Gemini models | + +## Sample Configuration + +```yaml +api_key: ${env.GEMINI_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_groq.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_groq.mdx new file mode 100644 index 0000000..1797035 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_groq.mdx @@ -0,0 +1,25 @@ +--- +description: "Groq inference provider for ultra-fast inference using Groq's LPU technology." +sidebar_label: Remote - Groq +title: remote::groq +--- + +# remote::groq + +## Description + +Groq inference provider for ultra-fast inference using Groq's LPU technology. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The Groq API key | +| `url` | `` | No | https://api.groq.com | The URL for the Groq AI server | + +## Sample Configuration + +```yaml +url: https://api.groq.com +api_key: ${env.GROQ_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_hf_endpoint.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_hf_endpoint.mdx new file mode 100644 index 0000000..771b24f --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_hf_endpoint.mdx @@ -0,0 +1,25 @@ +--- +description: "HuggingFace Inference Endpoints provider for dedicated model serving." +sidebar_label: Remote - Hf - Endpoint +title: remote::hf::endpoint +--- + +# remote::hf::endpoint + +## Description + +HuggingFace Inference Endpoints provider for dedicated model serving. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `endpoint_name` | `` | No | | The name of the Hugging Face Inference Endpoint in the format of '{namespace}/{endpoint_name}' (e.g. 'my-cool-org/meta-llama-3-1-8b-instruct-rce'). Namespace is optional and will default to the user account if not provided. | +| `api_token` | `pydantic.types.SecretStr \| None` | No | | Your Hugging Face user access token (will default to locally saved token if not provided) | + +## Sample Configuration + +```yaml +endpoint_name: ${env.INFERENCE_ENDPOINT_NAME} +api_token: ${env.HF_API_TOKEN} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_hf_serverless.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_hf_serverless.mdx new file mode 100644 index 0000000..1a89b8e --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_hf_serverless.mdx @@ -0,0 +1,25 @@ +--- +description: "HuggingFace Inference API serverless provider for on-demand model inference." +sidebar_label: Remote - Hf - Serverless +title: remote::hf::serverless +--- + +# remote::hf::serverless + +## Description + +HuggingFace Inference API serverless provider for on-demand model inference. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `huggingface_repo` | `` | No | | The model ID of the model on the Hugging Face Hub (e.g. 'meta-llama/Meta-Llama-3.1-70B-Instruct') | +| `api_token` | `pydantic.types.SecretStr \| None` | No | | Your Hugging Face user access token (will default to locally saved token if not provided) | + +## Sample Configuration + +```yaml +huggingface_repo: ${env.INFERENCE_MODEL} +api_token: ${env.HF_API_TOKEN} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_llama-openai-compat.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_llama-openai-compat.mdx new file mode 100644 index 0000000..cb624ad --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_llama-openai-compat.mdx @@ -0,0 +1,25 @@ +--- +description: "Llama OpenAI-compatible provider for using Llama models with OpenAI API format." +sidebar_label: Remote - Llama-Openai-Compat +title: remote::llama-openai-compat +--- + +# remote::llama-openai-compat + +## Description + +Llama OpenAI-compatible provider for using Llama models with OpenAI API format. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The Llama API key | +| `openai_compat_api_base` | `` | No | https://api.llama.com/compat/v1/ | The URL for the Llama API server | + +## Sample Configuration + +```yaml +openai_compat_api_base: https://api.llama.com/compat/v1/ +api_key: ${env.LLAMA_API_KEY} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_nvidia.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_nvidia.mdx new file mode 100644 index 0000000..4a8be5d --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_nvidia.mdx @@ -0,0 +1,28 @@ +--- +description: "NVIDIA inference provider for accessing NVIDIA NIM models and AI services." +sidebar_label: Remote - Nvidia +title: remote::nvidia +--- + +# remote::nvidia + +## Description + +NVIDIA inference provider for accessing NVIDIA NIM models and AI services. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | https://integrate.api.nvidia.com | A base url for accessing the NVIDIA NIM | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | The NVIDIA API key, only needed of using the hosted service | +| `timeout` | `` | No | 60 | Timeout for the HTTP requests | +| `append_api_version` | `` | No | True | When set to false, the API version will not be appended to the base_url. By default, it is true. | + +## Sample Configuration + +```yaml +url: ${env.NVIDIA_BASE_URL:=https://integrate.api.nvidia.com} +api_key: ${env.NVIDIA_API_KEY:=} +append_api_version: ${env.NVIDIA_APPEND_API_VERSION:=True} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_ollama.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_ollama.mdx new file mode 100644 index 0000000..5d9a4ad --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_ollama.mdx @@ -0,0 +1,24 @@ +--- +description: "Ollama inference provider for running local models through the Ollama runtime." +sidebar_label: Remote - Ollama +title: remote::ollama +--- + +# remote::ollama + +## Description + +Ollama inference provider for running local models through the Ollama runtime. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | http://localhost:11434 | | +| `refresh_models` | `` | No | False | Whether to refresh models periodically | + +## Sample Configuration + +```yaml +url: ${env.OLLAMA_URL:=http://localhost:11434} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_openai.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_openai.mdx new file mode 100644 index 0000000..56ca942 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_openai.mdx @@ -0,0 +1,25 @@ +--- +description: "OpenAI inference provider for accessing GPT models and other OpenAI services." +sidebar_label: Remote - Openai +title: remote::openai +--- + +# remote::openai + +## Description + +OpenAI inference provider for accessing GPT models and other OpenAI services. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | API key for OpenAI models | +| `base_url` | `` | No | https://api.openai.com/v1 | Base URL for OpenAI API | + +## Sample Configuration + +```yaml +api_key: ${env.OPENAI_API_KEY:=} +base_url: ${env.OPENAI_BASE_URL:=https://api.openai.com/v1} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_passthrough.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_passthrough.mdx new file mode 100644 index 0000000..972cc2a --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_passthrough.mdx @@ -0,0 +1,25 @@ +--- +description: "Passthrough inference provider for connecting to any external inference service not directly supported." +sidebar_label: Remote - Passthrough +title: remote::passthrough +--- + +# remote::passthrough + +## Description + +Passthrough inference provider for connecting to any external inference service not directly supported. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | | The URL for the passthrough endpoint | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | API Key for the passthrouth endpoint | + +## Sample Configuration + +```yaml +url: ${env.PASSTHROUGH_URL} +api_key: ${env.PASSTHROUGH_API_KEY} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_runpod.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_runpod.mdx new file mode 100644 index 0000000..2e8847d --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_runpod.mdx @@ -0,0 +1,25 @@ +--- +description: "RunPod inference provider for running models on RunPod's cloud GPU platform." +sidebar_label: Remote - Runpod +title: remote::runpod +--- + +# remote::runpod + +## Description + +RunPod inference provider for running models on RunPod's cloud GPU platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `str \| None` | No | | The URL for the Runpod model serving endpoint | +| `api_token` | `str \| None` | No | | The API token | + +## Sample Configuration + +```yaml +url: ${env.RUNPOD_URL:=} +api_token: ${env.RUNPOD_API_TOKEN} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_sambanova-openai-compat.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_sambanova-openai-compat.mdx new file mode 100644 index 0000000..9b4716d --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_sambanova-openai-compat.mdx @@ -0,0 +1,20 @@ +# remote::sambanova-openai-compat + +## Description + +SambaNova OpenAI-compatible provider for using SambaNova models with OpenAI API format. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The SambaNova API key | +| `openai_compat_api_base` | `` | No | https://api.sambanova.ai/v1 | The URL for the SambaNova API server | + +## Sample Configuration + +```yaml +openai_compat_api_base: https://api.sambanova.ai/v1 +api_key: ${env.SAMBANOVA_API_KEY:=} + +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_sambanova.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_sambanova.mdx new file mode 100644 index 0000000..6ee28b4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_sambanova.mdx @@ -0,0 +1,25 @@ +--- +description: "SambaNova inference provider for running models on SambaNova's dataflow architecture." +sidebar_label: Remote - Sambanova +title: remote::sambanova +--- + +# remote::sambanova + +## Description + +SambaNova inference provider for running models on SambaNova's dataflow architecture. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | https://api.sambanova.ai/v1 | The URL for the SambaNova AI server | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | The SambaNova cloud API Key | + +## Sample Configuration + +```yaml +url: https://api.sambanova.ai/v1 +api_key: ${env.SAMBANOVA_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_tgi.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_tgi.mdx new file mode 100644 index 0000000..3a34805 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_tgi.mdx @@ -0,0 +1,23 @@ +--- +description: "Text Generation Inference (TGI) provider for HuggingFace model serving." +sidebar_label: Remote - Tgi +title: remote::tgi +--- + +# remote::tgi + +## Description + +Text Generation Inference (TGI) provider for HuggingFace model serving. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | | The URL for the TGI serving endpoint | + +## Sample Configuration + +```yaml +url: ${env.TGI_URL:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_together.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_together.mdx new file mode 100644 index 0000000..da232a4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_together.mdx @@ -0,0 +1,26 @@ +--- +description: "Together AI inference provider for open-source models and collaborative AI development." +sidebar_label: Remote - Together +title: remote::together +--- + +# remote::together + +## Description + +Together AI inference provider for open-source models and collaborative AI development. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. | +| `url` | `` | No | https://api.together.xyz/v1 | The URL for the Together AI server | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | The Together AI API Key | + +## Sample Configuration + +```yaml +url: https://api.together.xyz/v1 +api_key: ${env.TOGETHER_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_vertexai.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_vertexai.mdx new file mode 100644 index 0000000..13a910d --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_vertexai.mdx @@ -0,0 +1,64 @@ +--- +description: | + Google Vertex AI inference provider enables you to use Google's Gemini models through Google Cloud's Vertex AI platform, providing several advantages: + + โ€ข Enterprise-grade security: Uses Google Cloud's security controls and IAM + โ€ข Better integration: Seamless integration with other Google Cloud services + โ€ข Advanced features: Access to additional Vertex AI features like model tuning and monitoring + โ€ข Authentication: Uses Google Cloud Application Default Credentials (ADC) instead of API keys + + Configuration: + - Set VERTEX_AI_PROJECT environment variable (required) + - Set VERTEX_AI_LOCATION environment variable (optional, defaults to us-central1) + - Use Google Cloud Application Default Credentials or service account key + + Authentication Setup: + Option 1 (Recommended): gcloud auth application-default login + Option 2: Set GOOGLE_APPLICATION_CREDENTIALS to service account key path + + Available Models: + - vertex_ai/gemini-2.0-flash + - vertex_ai/gemini-2.5-flash + - vertex_ai/gemini-2.5-pro +sidebar_label: Remote - Vertexai +title: remote::vertexai +--- + +# remote::vertexai + +## Description + +Google Vertex AI inference provider enables you to use Google's Gemini models through Google Cloud's Vertex AI platform, providing several advantages: + +โ€ข Enterprise-grade security: Uses Google Cloud's security controls and IAM +โ€ข Better integration: Seamless integration with other Google Cloud services +โ€ข Advanced features: Access to additional Vertex AI features like model tuning and monitoring +โ€ข Authentication: Uses Google Cloud Application Default Credentials (ADC) instead of API keys + +Configuration: +- Set VERTEX_AI_PROJECT environment variable (required) +- Set VERTEX_AI_LOCATION environment variable (optional, defaults to us-central1) +- Use Google Cloud Application Default Credentials or service account key + +Authentication Setup: +Option 1 (Recommended): gcloud auth application-default login +Option 2: Set GOOGLE_APPLICATION_CREDENTIALS to service account key path + +Available Models: +- vertex_ai/gemini-2.0-flash +- vertex_ai/gemini-2.5-flash +- vertex_ai/gemini-2.5-pro + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `project` | `` | No | | Google Cloud project ID for Vertex AI | +| `location` | `` | No | us-central1 | Google Cloud location for Vertex AI | + +## Sample Configuration + +```yaml +project: ${env.VERTEX_AI_PROJECT:=} +location: ${env.VERTEX_AI_LOCATION:=us-central1} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_vllm.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_vllm.mdx new file mode 100644 index 0000000..77b8e13 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_vllm.mdx @@ -0,0 +1,30 @@ +--- +description: "Remote vLLM inference provider for connecting to vLLM servers." +sidebar_label: Remote - Vllm +title: remote::vllm +--- + +# remote::vllm + +## Description + +Remote vLLM inference provider for connecting to vLLM servers. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `str \| None` | No | | The URL for the vLLM model serving endpoint | +| `max_tokens` | `` | No | 4096 | Maximum number of tokens to generate. | +| `api_token` | `str \| None` | No | fake | The API token | +| `tls_verify` | `bool \| str` | No | True | Whether to verify TLS certificates. Can be a boolean or a path to a CA certificate file. | +| `refresh_models` | `` | No | False | Whether to refresh models periodically | + +## Sample Configuration + +```yaml +url: ${env.VLLM_URL:=} +max_tokens: ${env.VLLM_MAX_TOKENS:=4096} +api_token: ${env.VLLM_API_TOKEN:=fake} +tls_verify: ${env.VLLM_TLS_VERIFY:=true} +``` diff --git a/versioned_docs/version-v0.2.23/providers/inference/remote_watsonx.mdx b/versioned_docs/version-v0.2.23/providers/inference/remote_watsonx.mdx new file mode 100644 index 0000000..1ceccc3 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/inference/remote_watsonx.mdx @@ -0,0 +1,28 @@ +--- +description: "IBM WatsonX inference provider for accessing AI models on IBM's WatsonX platform." +sidebar_label: Remote - Watsonx +title: remote::watsonx +--- + +# remote::watsonx + +## Description + +IBM WatsonX inference provider for accessing AI models on IBM's WatsonX platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | https://us-south.ml.cloud.ibm.com | A base url for accessing the watsonx.ai | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | The watsonx API key | +| `project_id` | `str \| None` | No | | The Project ID key | +| `timeout` | `` | No | 60 | Timeout for the HTTP requests | + +## Sample Configuration + +```yaml +url: ${env.WATSONX_BASE_URL:=https://us-south.ml.cloud.ibm.com} +api_key: ${env.WATSONX_API_KEY:=} +project_id: ${env.WATSONX_PROJECT_ID:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/openai.mdx b/versioned_docs/version-v0.2.23/providers/openai.mdx new file mode 100644 index 0000000..bcff587 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/openai.mdx @@ -0,0 +1,196 @@ +title: OpenAI Compatibility +description: OpenAI API Compatibility +sidebar_label: OpenAI Compatibility +sidebar_position: 1 +--- +## OpenAI API Compatibility + +### Server path + +Llama Stack exposes an OpenAI-compatible API endpoint at `/v1/openai/v1`. So, for a Llama Stack server running locally on port `8321`, the full url to the OpenAI-compatible API endpoint is `http://localhost:8321/v1/openai/v1`. + +### Clients + +You should be able to use any client that speaks OpenAI APIs with Llama Stack. We regularly test with the official Llama Stack clients as well as OpenAI's official Python client. + +#### Llama Stack Client + +When using the Llama Stack client, set the `base_url` to the root of your Llama Stack server. It will automatically route OpenAI-compatible requests to the right server endpoint for you. + +```python +from llama_stack_client import LlamaStackClient + +client = LlamaStackClient(base_url="http://localhost:8321") +``` + +#### OpenAI Client + +When using an OpenAI client, set the `base_url` to the `/v1/openai/v1` path on your Llama Stack server. + +```python +from openai import OpenAI + +client = OpenAI(base_url="http://localhost:8321/v1/openai/v1", api_key="none") +``` + +Regardless of the client you choose, the following code examples should all work the same. + +### APIs implemented + +#### Models + +Many of the APIs require you to pass in a model parameter. To see the list of models available in your Llama Stack server: + +```python +models = client.models.list() +``` + +#### Responses + +> **Note:** The Responses API implementation is still in active development. While it is quite usable, there are still unimplemented parts of the API. We'd love feedback on any use-cases you try that do not work to help prioritize the pieces left to implement. Please open issues in the [meta-llama/llama-stack](https://github.com/meta-llama/llama-stack) GitHub repository with details of anything that does not work. + +##### Simple inference + +Request: + +``` +response = client.responses.create( + model="meta-llama/Llama-3.2-3B-Instruct", + input="Write a haiku about coding." +) + +print(response.output_text) +``` +Example output: + +```text +Pixels dancing slow +Syntax whispers secrets sweet +Code's gentle silence +``` + +##### Structured Output + +Request: + +```python +response = client.responses.create( + model="meta-llama/Llama-3.2-3B-Instruct", + input=[ + { + "role": "system", + "content": "Extract the participants from the event information.", + }, + { + "role": "user", + "content": "Alice and Bob are going to a science fair on Friday.", + }, + ], + text={ + "format": { + "type": "json_schema", + "name": "participants", + "schema": { + "type": "object", + "properties": { + "participants": {"type": "array", "items": {"type": "string"}} + }, + "required": ["participants"], + }, + } + }, +) +print(response.output_text) +``` + +Example output: + +```text +{ "participants": ["Alice", "Bob"] } +``` + +#### Chat Completions + +##### Simple inference + +Request: + +```python +chat_completion = client.chat.completions.create( + model="meta-llama/Llama-3.2-3B-Instruct", + messages=[{"role": "user", "content": "Write a haiku about coding."}], +) + +print(chat_completion.choices[0].message.content) +``` + +Example output: + +```text +Lines of code unfold +Logic flows like a river +Code's gentle beauty +``` + +##### Structured Output + +Request: + +```python +chat_completion = client.chat.completions.create( + model="meta-llama/Llama-3.2-3B-Instruct", + messages=[ + { + "role": "system", + "content": "Extract the participants from the event information.", + }, + { + "role": "user", + "content": "Alice and Bob are going to a science fair on Friday.", + }, + ], + response_format={ + "type": "json_schema", + "json_schema": { + "name": "participants", + "schema": { + "type": "object", + "properties": { + "participants": {"type": "array", "items": {"type": "string"}} + }, + "required": ["participants"], + }, + }, + }, +) + +print(chat_completion.choices[0].message.content) +``` + +Example output: + +```text +{ "participants": ["Alice", "Bob"] } +``` + +#### Completions + +##### Simple inference + +Request: + +```python +completion = client.completions.create( + model="meta-llama/Llama-3.2-3B-Instruct", prompt="Write a haiku about coding." +) + +print(completion.choices[0].text) +``` + +Example output: + +```text +Lines of code unfurl +Logic whispers in the dark +Art in hidden form +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/index.mdx b/versioned_docs/version-v0.2.23/providers/post_training/index.mdx new file mode 100644 index 0000000..e3c8ba0 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Post Training +title: Post_Training +--- + +# Post_Training + +## Overview + +This section contains documentation for all available providers for the **post_training** API. diff --git a/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface-cpu.mdx b/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface-cpu.mdx new file mode 100644 index 0000000..4e64d57 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface-cpu.mdx @@ -0,0 +1,37 @@ +# inline::huggingface-cpu + +## Description + +HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `device` | `` | No | cuda | | +| `distributed_backend` | `Literal['fsdp', 'deepspeed'` | No | | | +| `checkpoint_format` | `Literal['full_state', 'huggingface'` | No | huggingface | | +| `chat_template` | `` | No | `<|user|>`
`{input}`
`<|assistant|>`
`{output}` | | +| `model_specific_config` | `` | No | `{'trust_remote_code': True, 'attn_implementation': 'sdpa'}` | | +| `max_seq_length` | `` | No | 2048 | | +| `gradient_checkpointing` | `` | No | False | | +| `save_total_limit` | `` | No | 3 | | +| `logging_steps` | `` | No | 10 | | +| `warmup_ratio` | `` | No | 0.1 | | +| `weight_decay` | `` | No | 0.01 | | +| `dataloader_num_workers` | `` | No | 4 | | +| `dataloader_pin_memory` | `` | No | True | | +| `dpo_beta` | `` | No | 0.1 | | +| `use_reference_model` | `` | No | True | | +| `dpo_loss_type` | `Literal['sigmoid', 'hinge', 'ipo', 'kto_pair'` | No | sigmoid | | +| `dpo_output_dir` | `` | No | | | + +## Sample Configuration + +```yaml +checkpoint_format: huggingface +distributed_backend: null +device: cpu +dpo_output_dir: ~/.llama/dummy/dpo_output + +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface-gpu.mdx b/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface-gpu.mdx new file mode 100644 index 0000000..ac7644d --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface-gpu.mdx @@ -0,0 +1,42 @@ +--- +description: "HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem." +sidebar_label: Huggingface-Gpu +title: inline::huggingface-gpu +--- + +# inline::huggingface-gpu + +## Description + +HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `device` | `` | No | cuda | | +| `distributed_backend` | `Literal['fsdp', 'deepspeed'` | No | | | +| `checkpoint_format` | `Literal['full_state', 'huggingface'` | No | huggingface | | +| `chat_template` | `` | No | `<|user|>`
`{input}`
`<|assistant|>`
`{output}` | | +| `model_specific_config` | `` | No | `{'trust_remote_code': True, 'attn_implementation': 'sdpa'}` | | +| `max_seq_length` | `` | No | 2048 | | +| `gradient_checkpointing` | `` | No | False | | +| `save_total_limit` | `` | No | 3 | | +| `logging_steps` | `` | No | 10 | | +| `warmup_ratio` | `` | No | 0.1 | | +| `weight_decay` | `` | No | 0.01 | | +| `dataloader_num_workers` | `` | No | 4 | | +| `dataloader_pin_memory` | `` | No | True | | +| `dpo_beta` | `` | No | 0.1 | | +| `use_reference_model` | `` | No | True | | +| `dpo_loss_type` | `Literal['sigmoid', 'hinge', 'ipo', 'kto_pair'` | No | sigmoid | | +| `dpo_output_dir` | `` | No | | | + +## Sample Configuration + +```yaml +checkpoint_format: huggingface +distributed_backend: null +device: cpu +dpo_output_dir: ~/.llama/dummy/dpo_output +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface.mdx b/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface.mdx new file mode 100644 index 0000000..870ff6e --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/inline_huggingface.mdx @@ -0,0 +1,37 @@ +# inline::huggingface + +## Description + +HuggingFace-based post-training provider for fine-tuning models using the HuggingFace ecosystem. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `device` | `` | No | cuda | | +| `distributed_backend` | `Literal['fsdp', 'deepspeed'` | No | | | +| `checkpoint_format` | `Literal['full_state', 'huggingface'` | No | huggingface | | +| `chat_template` | `` | No | `<|user|>`
`{input}`
`<|assistant|>`
`{output}` | | +| `model_specific_config` | `` | No | `{'trust_remote_code': True, 'attn_implementation': 'sdpa'}` | | +| `max_seq_length` | `` | No | 2048 | | +| `gradient_checkpointing` | `` | No | False | | +| `save_total_limit` | `` | No | 3 | | +| `logging_steps` | `` | No | 10 | | +| `warmup_ratio` | `` | No | 0.1 | | +| `weight_decay` | `` | No | 0.01 | | +| `dataloader_num_workers` | `` | No | 4 | | +| `dataloader_pin_memory` | `` | No | True | | +| `dpo_beta` | `` | No | 0.1 | | +| `use_reference_model` | `` | No | True | | +| `dpo_loss_type` | `Literal['sigmoid', 'hinge', 'ipo', 'kto_pair'` | No | sigmoid | | +| `dpo_output_dir` | `` | No | | | + +## Sample Configuration + +```yaml +checkpoint_format: huggingface +distributed_backend: null +device: cpu +dpo_output_dir: ~/.llama/dummy/dpo_output + +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune-cpu.mdx b/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune-cpu.mdx new file mode 100644 index 0000000..f789392 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune-cpu.mdx @@ -0,0 +1,24 @@ +--- +description: "TorchTune-based post-training provider for fine-tuning and optimizing models using Meta's TorchTune framework." +sidebar_label: Torchtune-Cpu +title: inline::torchtune-cpu +--- + +# inline::torchtune-cpu + +## Description + +TorchTune-based post-training provider for fine-tuning and optimizing models using Meta's TorchTune framework. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `torch_seed` | `int \| None` | No | | | +| `checkpoint_format` | `Literal['meta', 'huggingface'` | No | meta | | + +## Sample Configuration + +```yaml +checkpoint_format: meta +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune-gpu.mdx b/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune-gpu.mdx new file mode 100644 index 0000000..bd87797 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune-gpu.mdx @@ -0,0 +1,24 @@ +--- +description: "TorchTune-based post-training provider for fine-tuning and optimizing models using Meta's TorchTune framework." +sidebar_label: Torchtune-Gpu +title: inline::torchtune-gpu +--- + +# inline::torchtune-gpu + +## Description + +TorchTune-based post-training provider for fine-tuning and optimizing models using Meta's TorchTune framework. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `torch_seed` | `int \| None` | No | | | +| `checkpoint_format` | `Literal['meta', 'huggingface'` | No | meta | | + +## Sample Configuration + +```yaml +checkpoint_format: meta +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune.md b/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune.md new file mode 100644 index 0000000..2680951 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/inline_torchtune.md @@ -0,0 +1,19 @@ +# inline::torchtune + +## Description + +TorchTune-based post-training provider for fine-tuning and optimizing models using Meta's TorchTune framework. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `torch_seed` | `int \| None` | No | | | +| `checkpoint_format` | `Literal['meta', 'huggingface'` | No | meta | | + +## Sample Configuration + +```yaml +checkpoint_format: meta + +``` diff --git a/versioned_docs/version-v0.2.23/providers/post_training/remote_nvidia.mdx b/versioned_docs/version-v0.2.23/providers/post_training/remote_nvidia.mdx new file mode 100644 index 0000000..448ac4c --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/post_training/remote_nvidia.mdx @@ -0,0 +1,32 @@ +--- +description: "NVIDIA's post-training provider for fine-tuning models on NVIDIA's platform." +sidebar_label: Remote - Nvidia +title: remote::nvidia +--- + +# remote::nvidia + +## Description + +NVIDIA's post-training provider for fine-tuning models on NVIDIA's platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The NVIDIA API key. | +| `dataset_namespace` | `str \| None` | No | default | The NVIDIA dataset namespace. | +| `project_id` | `str \| None` | No | test-example-model@v1 | The NVIDIA project ID. | +| `customizer_url` | `str \| None` | No | | Base URL for the NeMo Customizer API | +| `timeout` | `` | No | 300 | Timeout for the NVIDIA Post Training API | +| `max_retries` | `` | No | 3 | Maximum number of retries for the NVIDIA Post Training API | +| `output_model_dir` | `` | No | test-example-model@v1 | Directory to save the output model | + +## Sample Configuration + +```yaml +api_key: ${env.NVIDIA_API_KEY:=} +dataset_namespace: ${env.NVIDIA_DATASET_NAMESPACE:=default} +project_id: ${env.NVIDIA_PROJECT_ID:=test-project} +customizer_url: ${env.NVIDIA_CUSTOMIZER_URL:=http://nemo.test} +``` diff --git a/versioned_docs/version-v0.2.23/providers/safety/index.mdx b/versioned_docs/version-v0.2.23/providers/safety/index.mdx new file mode 100644 index 0000000..3445b17 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Safety +title: Safety +--- + +# Safety + +## Overview + +This section contains documentation for all available providers for the **safety** API. diff --git a/versioned_docs/version-v0.2.23/providers/safety/inline_code-scanner.mdx b/versioned_docs/version-v0.2.23/providers/safety/inline_code-scanner.mdx new file mode 100644 index 0000000..3fc3c38 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/inline_code-scanner.mdx @@ -0,0 +1,17 @@ +--- +description: "Code Scanner safety provider for detecting security vulnerabilities and unsafe code patterns." +sidebar_label: Code-Scanner +title: inline::code-scanner +--- + +# inline::code-scanner + +## Description + +Code Scanner safety provider for detecting security vulnerabilities and unsafe code patterns. + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/safety/inline_llama-guard.mdx b/versioned_docs/version-v0.2.23/providers/safety/inline_llama-guard.mdx new file mode 100644 index 0000000..65866c9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/inline_llama-guard.mdx @@ -0,0 +1,23 @@ +--- +description: "Llama Guard safety provider for content moderation and safety filtering using Meta's Llama Guard model." +sidebar_label: Llama-Guard +title: inline::llama-guard +--- + +# inline::llama-guard + +## Description + +Llama Guard safety provider for content moderation and safety filtering using Meta's Llama Guard model. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `excluded_categories` | `list[str` | No | [] | | + +## Sample Configuration + +```yaml +excluded_categories: [] +``` diff --git a/versioned_docs/version-v0.2.23/providers/safety/inline_prompt-guard.mdx b/versioned_docs/version-v0.2.23/providers/safety/inline_prompt-guard.mdx new file mode 100644 index 0000000..c52e03e --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/inline_prompt-guard.mdx @@ -0,0 +1,23 @@ +--- +description: "Prompt Guard safety provider for detecting and filtering unsafe prompts and content." +sidebar_label: Prompt-Guard +title: inline::prompt-guard +--- + +# inline::prompt-guard + +## Description + +Prompt Guard safety provider for detecting and filtering unsafe prompts and content. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `guard_type` | `` | No | injection | | + +## Sample Configuration + +```yaml +guard_type: injection +``` diff --git a/versioned_docs/version-v0.2.23/providers/safety/remote_bedrock.mdx b/versioned_docs/version-v0.2.23/providers/safety/remote_bedrock.mdx new file mode 100644 index 0000000..5461d7c --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/remote_bedrock.mdx @@ -0,0 +1,32 @@ +--- +description: "AWS Bedrock safety provider for content moderation using AWS's safety services." +sidebar_label: Remote - Bedrock +title: remote::bedrock +--- + +# remote::bedrock + +## Description + +AWS Bedrock safety provider for content moderation using AWS's safety services. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `aws_access_key_id` | `str \| None` | No | | The AWS access key to use. Default use environment variable: AWS_ACCESS_KEY_ID | +| `aws_secret_access_key` | `str \| None` | No | | The AWS secret access key to use. Default use environment variable: AWS_SECRET_ACCESS_KEY | +| `aws_session_token` | `str \| None` | No | | The AWS session token to use. Default use environment variable: AWS_SESSION_TOKEN | +| `region_name` | `str \| None` | No | | The default AWS Region to use, for example, us-west-1 or us-west-2.Default use environment variable: AWS_DEFAULT_REGION | +| `profile_name` | `str \| None` | No | | The profile name that contains credentials to use.Default use environment variable: AWS_PROFILE | +| `total_max_attempts` | `int \| None` | No | | An integer representing the maximum number of attempts that will be made for a single request, including the initial attempt. Default use environment variable: AWS_MAX_ATTEMPTS | +| `retry_mode` | `str \| None` | No | | A string representing the type of retries Boto3 will perform.Default use environment variable: AWS_RETRY_MODE | +| `connect_timeout` | `float \| None` | No | 60.0 | The time in seconds till a timeout exception is thrown when attempting to make a connection. The default is 60 seconds. | +| `read_timeout` | `float \| None` | No | 60.0 | The time in seconds till a timeout exception is thrown when attempting to read from a connection.The default is 60 seconds. | +| `session_ttl` | `int \| None` | No | 3600 | The time in seconds till a session expires. The default is 3600 seconds (1 hour). | + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/safety/remote_nvidia.mdx b/versioned_docs/version-v0.2.23/providers/safety/remote_nvidia.mdx new file mode 100644 index 0000000..0f665e6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/remote_nvidia.mdx @@ -0,0 +1,25 @@ +--- +description: "NVIDIA's safety provider for content moderation and safety filtering." +sidebar_label: Remote - Nvidia +title: remote::nvidia +--- + +# remote::nvidia + +## Description + +NVIDIA's safety provider for content moderation and safety filtering. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `guardrails_service_url` | `` | No | http://0.0.0.0:7331 | The url for accessing the Guardrails service | +| `config_id` | `str \| None` | No | self-check | Guardrails configuration ID to use from the Guardrails configuration store | + +## Sample Configuration + +```yaml +guardrails_service_url: ${env.GUARDRAILS_SERVICE_URL:=http://localhost:7331} +config_id: ${env.NVIDIA_GUARDRAILS_CONFIG_ID:=self-check} +``` diff --git a/versioned_docs/version-v0.2.23/providers/safety/remote_sambanova.mdx b/versioned_docs/version-v0.2.23/providers/safety/remote_sambanova.mdx new file mode 100644 index 0000000..da70fce --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/safety/remote_sambanova.mdx @@ -0,0 +1,25 @@ +--- +description: "SambaNova's safety provider for content moderation and safety filtering." +sidebar_label: Remote - Sambanova +title: remote::sambanova +--- + +# remote::sambanova + +## Description + +SambaNova's safety provider for content moderation and safety filtering. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `` | No | https://api.sambanova.ai/v1 | The URL for the SambaNova AI server | +| `api_key` | `pydantic.types.SecretStr \| None` | No | | The SambaNova cloud API Key | + +## Sample Configuration + +```yaml +url: https://api.sambanova.ai/v1 +api_key: ${env.SAMBANOVA_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/scoring/index.mdx b/versioned_docs/version-v0.2.23/providers/scoring/index.mdx new file mode 100644 index 0000000..41d63b4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/scoring/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Scoring +title: Scoring +--- + +# Scoring + +## Overview + +This section contains documentation for all available providers for the **scoring** API. diff --git a/versioned_docs/version-v0.2.23/providers/scoring/inline_basic.mdx b/versioned_docs/version-v0.2.23/providers/scoring/inline_basic.mdx new file mode 100644 index 0000000..cbafbc4 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/scoring/inline_basic.mdx @@ -0,0 +1,17 @@ +--- +description: "Basic scoring provider for simple evaluation metrics and scoring functions." +sidebar_label: Basic +title: inline::basic +--- + +# inline::basic + +## Description + +Basic scoring provider for simple evaluation metrics and scoring functions. + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/scoring/inline_braintrust.mdx b/versioned_docs/version-v0.2.23/providers/scoring/inline_braintrust.mdx new file mode 100644 index 0000000..d12f9de --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/scoring/inline_braintrust.mdx @@ -0,0 +1,23 @@ +--- +description: "Braintrust scoring provider for evaluation and scoring using the Braintrust platform." +sidebar_label: Braintrust +title: inline::braintrust +--- + +# inline::braintrust + +## Description + +Braintrust scoring provider for evaluation and scoring using the Braintrust platform. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `openai_api_key` | `str \| None` | No | | The OpenAI API Key | + +## Sample Configuration + +```yaml +openai_api_key: ${env.OPENAI_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/scoring/inline_llm-as-judge.mdx b/versioned_docs/version-v0.2.23/providers/scoring/inline_llm-as-judge.mdx new file mode 100644 index 0000000..22f3266 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/scoring/inline_llm-as-judge.mdx @@ -0,0 +1,17 @@ +--- +description: "LLM-as-judge scoring provider that uses language models to evaluate and score responses." +sidebar_label: Llm-As-Judge +title: inline::llm-as-judge +--- + +# inline::llm-as-judge + +## Description + +LLM-as-judge scoring provider that uses language models to evaluate and score responses. + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/telemetry/index.mdx b/versioned_docs/version-v0.2.23/providers/telemetry/index.mdx new file mode 100644 index 0000000..07190d6 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/telemetry/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Telemetry +title: Telemetry +--- + +# Telemetry + +## Overview + +This section contains documentation for all available providers for the **telemetry** API. diff --git a/versioned_docs/version-v0.2.23/providers/telemetry/inline_meta-reference.mdx b/versioned_docs/version-v0.2.23/providers/telemetry/inline_meta-reference.mdx new file mode 100644 index 0000000..13fab87 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/telemetry/inline_meta-reference.mdx @@ -0,0 +1,29 @@ +--- +description: "Meta's reference implementation of telemetry and observability using OpenTelemetry." +sidebar_label: Meta-Reference +title: inline::meta-reference +--- + +# inline::meta-reference + +## Description + +Meta's reference implementation of telemetry and observability using OpenTelemetry. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `otel_exporter_otlp_endpoint` | `str \| None` | No | | The OpenTelemetry collector endpoint URL (base URL for traces, metrics, and logs). If not set, the SDK will use OTEL_EXPORTER_OTLP_ENDPOINT environment variable. | +| `service_name` | `` | No | โ€‹ | The service name to use for telemetry | +| `sinks` | `list[inline.telemetry.meta_reference.config.TelemetrySink` | No | [<TelemetrySink.CONSOLE: 'console'>, <TelemetrySink.SQLITE: 'sqlite'>] | List of telemetry sinks to enable (possible values: otel_trace, otel_metric, sqlite, console) | +| `sqlite_db_path` | `` | No | ~/.llama/runtime/trace_store.db | The path to the SQLite database to use for storing traces | + +## Sample Configuration + +```yaml +service_name: "${env.OTEL_SERVICE_NAME:=\u200B}" +sinks: ${env.TELEMETRY_SINKS:=console,sqlite} +sqlite_db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/trace_store.db +otel_exporter_otlp_endpoint: ${env.OTEL_EXPORTER_OTLP_ENDPOINT:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/index.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/index.mdx new file mode 100644 index 0000000..ab50509 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Tool Runtime +title: Tool_Runtime +--- + +# Tool_Runtime + +## Overview + +This section contains documentation for all available providers for the **tool_runtime** API. diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/inline_rag-runtime.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/inline_rag-runtime.mdx new file mode 100644 index 0000000..97428c2 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/inline_rag-runtime.mdx @@ -0,0 +1,17 @@ +--- +description: "RAG (Retrieval-Augmented Generation) tool runtime for document ingestion, chunking, and semantic search." +sidebar_label: Rag-Runtime +title: inline::rag-runtime +--- + +# inline::rag-runtime + +## Description + +RAG (Retrieval-Augmented Generation) tool runtime for document ingestion, chunking, and semantic search. + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_bing-search.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_bing-search.mdx new file mode 100644 index 0000000..ec06bc2 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_bing-search.mdx @@ -0,0 +1,24 @@ +--- +description: "Bing Search tool for web search capabilities using Microsoft's search engine." +sidebar_label: Remote - Bing-Search +title: remote::bing-search +--- + +# remote::bing-search + +## Description + +Bing Search tool for web search capabilities using Microsoft's search engine. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | | +| `top_k` | `` | No | 3 | | + +## Sample Configuration + +```yaml +api_key: ${env.BING_API_KEY:} +``` diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_brave-search.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_brave-search.mdx new file mode 100644 index 0000000..3aeed67 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_brave-search.mdx @@ -0,0 +1,25 @@ +--- +description: "Brave Search tool for web search capabilities with privacy-focused results." +sidebar_label: Remote - Brave-Search +title: remote::brave-search +--- + +# remote::brave-search + +## Description + +Brave Search tool for web search capabilities with privacy-focused results. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The Brave Search API Key | +| `max_results` | `` | No | 3 | The maximum number of results to return | + +## Sample Configuration + +```yaml +api_key: ${env.BRAVE_SEARCH_API_KEY:=} +max_results: 3 +``` diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_model-context-protocol.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_model-context-protocol.mdx new file mode 100644 index 0000000..869ca27 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_model-context-protocol.mdx @@ -0,0 +1,17 @@ +--- +description: "Model Context Protocol (MCP) tool for standardized tool calling and context management." +sidebar_label: Remote - Model-Context-Protocol +title: remote::model-context-protocol +--- + +# remote::model-context-protocol + +## Description + +Model Context Protocol (MCP) tool for standardized tool calling and context management. + +## Sample Configuration + +```yaml +{} +``` diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_tavily-search.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_tavily-search.mdx new file mode 100644 index 0000000..fdca31b --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_tavily-search.mdx @@ -0,0 +1,25 @@ +--- +description: "Tavily Search tool for AI-optimized web search with structured results." +sidebar_label: Remote - Tavily-Search +title: remote::tavily-search +--- + +# remote::tavily-search + +## Description + +Tavily Search tool for AI-optimized web search with structured results. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | The Tavily Search API Key | +| `max_results` | `` | No | 3 | The maximum number of results to return | + +## Sample Configuration + +```yaml +api_key: ${env.TAVILY_SEARCH_API_KEY:=} +max_results: 3 +``` diff --git a/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_wolfram-alpha.mdx b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_wolfram-alpha.mdx new file mode 100644 index 0000000..96bc417 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/tool_runtime/remote_wolfram-alpha.mdx @@ -0,0 +1,23 @@ +--- +description: "Wolfram Alpha tool for computational knowledge and mathematical calculations." +sidebar_label: Remote - Wolfram-Alpha +title: remote::wolfram-alpha +--- + +# remote::wolfram-alpha + +## Description + +Wolfram Alpha tool for computational knowledge and mathematical calculations. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `api_key` | `str \| None` | No | | | + +## Sample Configuration + +```yaml +api_key: ${env.WOLFRAM_ALPHA_API_KEY:=} +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/index.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/index.mdx new file mode 100644 index 0000000..4c4c81e --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/index.mdx @@ -0,0 +1,10 @@ +--- +sidebar_label: Vector Io +title: Vector_Io +--- + +# Vector_Io + +## Overview + +This section contains documentation for all available providers for the **vector_io** API. diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_chromadb.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_chromadb.mdx new file mode 100644 index 0000000..a1858ea --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_chromadb.mdx @@ -0,0 +1,91 @@ +--- +description: | + [Chroma](https://www.trychroma.com/) is an inline and remote vector + database provider for Llama Stack. It allows you to store and query vectors directly within a Chroma database. + That means you're not limited to storing vectors in memory or in a separate service. + + ## Features + Chroma supports: + - Store embeddings and their metadata + - Vector search + - Full-text search + - Document storage + - Metadata filtering + - Multi-modal retrieval + + ## Usage + + To use Chrome in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use chroma. + 3. Start storing and querying vectors. + + ## Installation + + You can install chroma using pip: + + ```bash + pip install chromadb + ``` + + ## Documentation + See [Chroma's documentation](https://docs.trychroma.com/docs/overview/introduction) for more details about Chroma in general. +sidebar_label: Chromadb +title: inline::chromadb +--- + +# inline::chromadb + +## Description + + +[Chroma](https://www.trychroma.com/) is an inline and remote vector +database provider for Llama Stack. It allows you to store and query vectors directly within a Chroma database. +That means you're not limited to storing vectors in memory or in a separate service. + +## Features +Chroma supports: +- Store embeddings and their metadata +- Vector search +- Full-text search +- Document storage +- Metadata filtering +- Multi-modal retrieval + +## Usage + +To use Chrome in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use chroma. +3. Start storing and querying vectors. + +## Installation + +You can install chroma using pip: + +```bash +pip install chromadb +``` + +## Documentation +See [Chroma's documentation](https://docs.trychroma.com/docs/overview/introduction) for more details about Chroma in general. + + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `db_path` | `` | No | | | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Config for KV store backend | + +## Sample Configuration + +```yaml +db_path: ${env.CHROMADB_PATH} +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/chroma_inline_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_faiss.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_faiss.mdx new file mode 100644 index 0000000..03bc2a9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_faiss.mdx @@ -0,0 +1,106 @@ +--- +description: | + [Faiss](https://github.com/facebookresearch/faiss) is an inline vector database provider for Llama Stack. It + allows you to store and query vectors directly in memory. + That means you'll get fast and efficient vector retrieval. + + ## Features + + - Lightweight and easy to use + - Fully integrated with Llama Stack + - GPU support + - **Vector search** - FAISS supports pure vector similarity search using embeddings + + ## Search Modes + + **Supported:** + - **Vector Search** (`mode="vector"`): Performs vector similarity search using embeddings + + **Not Supported:** + - **Keyword Search** (`mode="keyword"`): Not supported by FAISS + - **Hybrid Search** (`mode="hybrid"`): Not supported by FAISS + + > **Note**: FAISS is designed as a pure vector similarity search library. See the [FAISS GitHub repository](https://github.com/facebookresearch/faiss) for more details about FAISS's core functionality. + + ## Usage + + To use Faiss in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use Faiss. + 3. Start storing and querying vectors. + + ## Installation + + You can install Faiss using pip: + + ```bash + pip install faiss-cpu + ``` + ## Documentation + See [Faiss' documentation](https://faiss.ai/) or the [Faiss Wiki](https://github.com/facebookresearch/faiss/wiki) for + more details about Faiss in general. +sidebar_label: Faiss +title: inline::faiss +--- + +# inline::faiss + +## Description + + +[Faiss](https://github.com/facebookresearch/faiss) is an inline vector database provider for Llama Stack. It +allows you to store and query vectors directly in memory. +That means you'll get fast and efficient vector retrieval. + +## Features + +- Lightweight and easy to use +- Fully integrated with Llama Stack +- GPU support +- **Vector search** - FAISS supports pure vector similarity search using embeddings + +## Search Modes + +**Supported:** +- **Vector Search** (`mode="vector"`): Performs vector similarity search using embeddings + +**Not Supported:** +- **Keyword Search** (`mode="keyword"`): Not supported by FAISS +- **Hybrid Search** (`mode="hybrid"`): Not supported by FAISS + +> **Note**: FAISS is designed as a pure vector similarity search library. See the [FAISS GitHub repository](https://github.com/facebookresearch/faiss) for more details about FAISS's core functionality. + +## Usage + +To use Faiss in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use Faiss. +3. Start storing and querying vectors. + +## Installation + +You can install Faiss using pip: + +```bash +pip install faiss-cpu +``` +## Documentation +See [Faiss' documentation](https://faiss.ai/) or the [Faiss Wiki](https://github.com/facebookresearch/faiss/wiki) for +more details about Faiss in general. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/faiss_store.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_meta-reference.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_meta-reference.mdx new file mode 100644 index 0000000..bcad867 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_meta-reference.mdx @@ -0,0 +1,30 @@ +--- +description: "Meta's reference implementation of a vector database." +sidebar_label: Meta-Reference +title: inline::meta-reference +--- + +# inline::meta-reference + +## Description + +Meta's reference implementation of a vector database. + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/faiss_store.db +``` +## Deprecation Notice + +:::warning +Please use the `inline::faiss` provider instead. +::: diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_milvus.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_milvus.mdx new file mode 100644 index 0000000..7e6f15c --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_milvus.mdx @@ -0,0 +1,30 @@ +--- +description: "Please refer to the remote provider documentation." +sidebar_label: Milvus +title: inline::milvus +--- + +# inline::milvus + +## Description + + +Please refer to the remote provider documentation. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `db_path` | `` | No | | | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Config for KV store backend (SQLite only for now) | +| `consistency_level` | `` | No | Strong | The consistency level of the Milvus server | + +## Sample Configuration + +```yaml +db_path: ${env.MILVUS_DB_PATH:=~/.llama/dummy}/milvus.db +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/milvus_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_qdrant.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_qdrant.mdx new file mode 100644 index 0000000..5c9ab10 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_qdrant.mdx @@ -0,0 +1,110 @@ +--- +description: | + [Qdrant](https://qdrant.tech/documentation/) is an inline and remote vector database provider for Llama Stack. It + allows you to store and query vectors directly in memory. + That means you'll get fast and efficient vector retrieval. + + > By default, Qdrant stores vectors in RAM, delivering incredibly fast access for datasets that fit comfortably in + > memory. But when your dataset exceeds RAM capacity, Qdrant offers Memmap as an alternative. + > + > \[[An Introduction to Vector Databases](https://qdrant.tech/articles/what-is-a-vector-database/)\] + + + + ## Features + + - Lightweight and easy to use + - Fully integrated with Llama Stack + - Apache 2.0 license terms + - Store embeddings and their metadata + - Supports search by + [Keyword](https://qdrant.tech/articles/qdrant-introduces-full-text-filters-and-indexes/) + and [Hybrid](https://qdrant.tech/articles/hybrid-search/#building-a-hybrid-search-system-in-qdrant) search + - [Multilingual and Multimodal retrieval](https://qdrant.tech/documentation/multimodal-search/) + - [Medatata filtering](https://qdrant.tech/articles/vector-search-filtering/) + - [GPU support](https://qdrant.tech/documentation/guides/running-with-gpu/) + + ## Usage + + To use Qdrant in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use Qdrant. + 3. Start storing and querying vectors. + + ## Installation + + You can install Qdrant using docker: + + ```bash + docker pull qdrant/qdrant + ``` + ## Documentation + See the [Qdrant documentation](https://qdrant.tech/documentation/) for more details about Qdrant in general. +sidebar_label: Qdrant +title: inline::qdrant +--- + +# inline::qdrant + +## Description + + +[Qdrant](https://qdrant.tech/documentation/) is an inline and remote vector database provider for Llama Stack. It +allows you to store and query vectors directly in memory. +That means you'll get fast and efficient vector retrieval. + +> By default, Qdrant stores vectors in RAM, delivering incredibly fast access for datasets that fit comfortably in +> memory. But when your dataset exceeds RAM capacity, Qdrant offers Memmap as an alternative. +> +> \[[An Introduction to Vector Databases](https://qdrant.tech/articles/what-is-a-vector-database/)\] + + + +## Features + +- Lightweight and easy to use +- Fully integrated with Llama Stack +- Apache 2.0 license terms +- Store embeddings and their metadata +- Supports search by + [Keyword](https://qdrant.tech/articles/qdrant-introduces-full-text-filters-and-indexes/) + and [Hybrid](https://qdrant.tech/articles/hybrid-search/#building-a-hybrid-search-system-in-qdrant) search +- [Multilingual and Multimodal retrieval](https://qdrant.tech/documentation/multimodal-search/) +- [Medatata filtering](https://qdrant.tech/articles/vector-search-filtering/) +- [GPU support](https://qdrant.tech/documentation/guides/running-with-gpu/) + +## Usage + +To use Qdrant in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use Qdrant. +3. Start storing and querying vectors. + +## Installation + +You can install Qdrant using docker: + +```bash +docker pull qdrant/qdrant +``` +## Documentation +See the [Qdrant documentation](https://qdrant.tech/documentation/) for more details about Qdrant in general. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `path` | `` | No | | | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +path: ${env.QDRANT_PATH:=~/.llama/~/.llama/dummy}/qdrant.db +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/qdrant_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_sqlite-vec.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_sqlite-vec.mdx new file mode 100644 index 0000000..aa6992a --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_sqlite-vec.mdx @@ -0,0 +1,420 @@ +--- +description: | + [SQLite-Vec](https://github.com/asg017/sqlite-vec) is an inline vector database provider for Llama Stack. It + allows you to store and query vectors directly within an SQLite database. + That means you're not limited to storing vectors in memory or in a separate service. + + ## Features + + - Lightweight and easy to use + - Fully integrated with Llama Stacks + - Uses disk-based storage for persistence, allowing for larger vector storage + + ### Comparison to Faiss + + The choice between Faiss and sqlite-vec should be made based on the needs of your application, + as they have different strengths. + + #### Choosing the Right Provider + + Scenario | Recommended Tool | Reason + -- |-----------------| -- + Online Analytical Processing (OLAP) | Faiss | Fast, in-memory searches + Online Transaction Processing (OLTP) | sqlite-vec | Frequent writes and reads + Frequent writes | sqlite-vec | Efficient disk-based storage and incremental indexing + Large datasets | sqlite-vec | Disk-based storage for larger vector storage + Datasets that can fit in memory, frequent reads | Faiss | Optimized for speed, indexing, and GPU acceleration + + #### Empirical Example + + Consider the histogram below in which 10,000 randomly generated strings were inserted + in batches of 100 into both Faiss and sqlite-vec using `client.tool_runtime.rag_tool.insert()`. + + ```{image} ../../../../_static/providers/vector_io/write_time_comparison_sqlite-vec-faiss.png + :alt: Comparison of SQLite-Vec and Faiss write times + :width: 400px + ``` + + You will notice that the average write time for `sqlite-vec` was 788ms, compared to + 47,640ms for Faiss. While the number is jarring, if you look at the distribution, you can see that it is rather + uniformly spread across the [1500, 100000] interval. + + Looking at each individual write in the order that the documents are inserted you'll see the increase in + write speed as Faiss reindexes the vectors after each write. + ```{image} ../../../../_static/providers/vector_io/write_time_sequence_sqlite-vec-faiss.png + :alt: Comparison of SQLite-Vec and Faiss write times + :width: 400px + ``` + + In comparison, the read times for Faiss was on average 10% faster than sqlite-vec. + The modes of the two distributions highlight the differences much further where Faiss + will likely yield faster read performance. + + ```{image} ../../../../_static/providers/vector_io/read_time_comparison_sqlite-vec-faiss.png + :alt: Comparison of SQLite-Vec and Faiss read times + :width: 400px + ``` + + ## Usage + + To use sqlite-vec in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use SQLite-Vec. + 3. Start storing and querying vectors. + + The SQLite-vec provider supports three search modes: + + 1. **Vector Search** (`mode="vector"`): Performs pure vector similarity search using the embeddings. + 2. **Keyword Search** (`mode="keyword"`): Performs full-text search using SQLite's FTS5. + 3. **Hybrid Search** (`mode="hybrid"`): Combines both vector and keyword search for better results. First performs keyword search to get candidate matches, then applies vector similarity search on those candidates. + + Example with hybrid search: + ```python + response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={"mode": "hybrid", "max_chunks": 3, "score_threshold": 0.7}, + ) + + # Using RRF ranker + response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={ + "mode": "hybrid", + "max_chunks": 3, + "score_threshold": 0.7, + "ranker": {"type": "rrf", "impact_factor": 60.0}, + }, + ) + + # Using weighted ranker + response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={ + "mode": "hybrid", + "max_chunks": 3, + "score_threshold": 0.7, + "ranker": {"type": "weighted", "alpha": 0.7}, # 70% vector, 30% keyword + }, + ) + ``` + + Example with explicit vector search: + ```python + response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={"mode": "vector", "max_chunks": 3, "score_threshold": 0.7}, + ) + ``` + + Example with keyword search: + ```python + response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={"mode": "keyword", "max_chunks": 3, "score_threshold": 0.7}, + ) + ``` + + ## Supported Search Modes + + The SQLite vector store supports three search modes: + + 1. **Vector Search** (`mode="vector"`): Uses vector similarity to find relevant chunks + 2. **Keyword Search** (`mode="keyword"`): Uses keyword matching to find relevant chunks + 3. **Hybrid Search** (`mode="hybrid"`): Combines both vector and keyword scores using a ranker + + ### Hybrid Search + + Hybrid search combines the strengths of both vector and keyword search by: + - Computing vector similarity scores + - Computing keyword match scores + - Using a ranker to combine these scores + + Two ranker types are supported: + + 1. **RRF (Reciprocal Rank Fusion)**: + - Combines ranks from both vector and keyword results + - Uses an impact factor (default: 60.0) to control the weight of higher-ranked results + - Good for balancing between vector and keyword results + - The default impact factor of 60.0 comes from the original RRF paper by Cormack et al. (2009) [^1], which found this value to provide optimal performance across various retrieval tasks + + 2. **Weighted**: + - Linearly combines normalized vector and keyword scores + - Uses an alpha parameter (0-1) to control the blend: + - alpha=0: Only use keyword scores + - alpha=1: Only use vector scores + - alpha=0.5: Equal weight to both (default) + + Example using RAGQueryConfig with different search modes: + + ```python + from llama_stack.apis.tools import RAGQueryConfig, RRFRanker, WeightedRanker + + # Vector search + config = RAGQueryConfig(mode="vector", max_chunks=5) + + # Keyword search + config = RAGQueryConfig(mode="keyword", max_chunks=5) + + # Hybrid search with custom RRF ranker + config = RAGQueryConfig( + mode="hybrid", + max_chunks=5, + ranker=RRFRanker(impact_factor=50.0), # Custom impact factor + ) + + # Hybrid search with weighted ranker + config = RAGQueryConfig( + mode="hybrid", + max_chunks=5, + ranker=WeightedRanker(alpha=0.7), # 70% vector, 30% keyword + ) + + # Hybrid search with default RRF ranker + config = RAGQueryConfig( + mode="hybrid", max_chunks=5 + ) # Will use RRF with impact_factor=60.0 + ``` + + Note: The ranker configuration is only used in hybrid mode. For vector or keyword modes, the ranker parameter is ignored. + + ## Installation + + You can install SQLite-Vec using pip: + + ```bash + pip install sqlite-vec + ``` + + ## Documentation + + See [sqlite-vec's GitHub repo](https://github.com/asg017/sqlite-vec/tree/main) for more details about sqlite-vec in general. + + [^1]: Cormack, G. V., Clarke, C. L., & Buettcher, S. (2009). [Reciprocal rank fusion outperforms condorcet and individual rank learning methods](https://dl.acm.org/doi/10.1145/1571941.1572114). In Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval (pp. 758-759). +sidebar_label: Sqlite-Vec +title: inline::sqlite-vec +--- + +# inline::sqlite-vec + +## Description + + +[SQLite-Vec](https://github.com/asg017/sqlite-vec) is an inline vector database provider for Llama Stack. It +allows you to store and query vectors directly within an SQLite database. +That means you're not limited to storing vectors in memory or in a separate service. + +## Features + +- Lightweight and easy to use +- Fully integrated with Llama Stacks +- Uses disk-based storage for persistence, allowing for larger vector storage + +### Comparison to Faiss + +The choice between Faiss and sqlite-vec should be made based on the needs of your application, +as they have different strengths. + +#### Choosing the Right Provider + +Scenario | Recommended Tool | Reason +-- |-----------------| -- +Online Analytical Processing (OLAP) | Faiss | Fast, in-memory searches +Online Transaction Processing (OLTP) | sqlite-vec | Frequent writes and reads +Frequent writes | sqlite-vec | Efficient disk-based storage and incremental indexing +Large datasets | sqlite-vec | Disk-based storage for larger vector storage +Datasets that can fit in memory, frequent reads | Faiss | Optimized for speed, indexing, and GPU acceleration + +#### Empirical Example + +Consider the histogram below in which 10,000 randomly generated strings were inserted +in batches of 100 into both Faiss and sqlite-vec using `client.tool_runtime.rag_tool.insert()`. + +```{image} ../../../../_static/providers/vector_io/write_time_comparison_sqlite-vec-faiss.png +:alt: Comparison of SQLite-Vec and Faiss write times +:width: 400px +``` + +You will notice that the average write time for `sqlite-vec` was 788ms, compared to +47,640ms for Faiss. While the number is jarring, if you look at the distribution, you can see that it is rather +uniformly spread across the [1500, 100000] interval. + +Looking at each individual write in the order that the documents are inserted you'll see the increase in +write speed as Faiss reindexes the vectors after each write. +```{image} ../../../../_static/providers/vector_io/write_time_sequence_sqlite-vec-faiss.png +:alt: Comparison of SQLite-Vec and Faiss write times +:width: 400px +``` + +In comparison, the read times for Faiss was on average 10% faster than sqlite-vec. +The modes of the two distributions highlight the differences much further where Faiss +will likely yield faster read performance. + +```{image} ../../../../_static/providers/vector_io/read_time_comparison_sqlite-vec-faiss.png +:alt: Comparison of SQLite-Vec and Faiss read times +:width: 400px +``` + +## Usage + +To use sqlite-vec in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use SQLite-Vec. +3. Start storing and querying vectors. + +The SQLite-vec provider supports three search modes: + +1. **Vector Search** (`mode="vector"`): Performs pure vector similarity search using the embeddings. +2. **Keyword Search** (`mode="keyword"`): Performs full-text search using SQLite's FTS5. +3. **Hybrid Search** (`mode="hybrid"`): Combines both vector and keyword search for better results. First performs keyword search to get candidate matches, then applies vector similarity search on those candidates. + +Example with hybrid search: +```python +response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={"mode": "hybrid", "max_chunks": 3, "score_threshold": 0.7}, +) + +# Using RRF ranker +response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={ + "mode": "hybrid", + "max_chunks": 3, + "score_threshold": 0.7, + "ranker": {"type": "rrf", "impact_factor": 60.0}, + }, +) + +# Using weighted ranker +response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={ + "mode": "hybrid", + "max_chunks": 3, + "score_threshold": 0.7, + "ranker": {"type": "weighted", "alpha": 0.7}, # 70% vector, 30% keyword + }, +) +``` + +Example with explicit vector search: +```python +response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={"mode": "vector", "max_chunks": 3, "score_threshold": 0.7}, +) +``` + +Example with keyword search: +```python +response = await vector_io.query_chunks( + vector_db_id="my_db", + query="your query here", + params={"mode": "keyword", "max_chunks": 3, "score_threshold": 0.7}, +) +``` + +## Supported Search Modes + +The SQLite vector store supports three search modes: + +1. **Vector Search** (`mode="vector"`): Uses vector similarity to find relevant chunks +2. **Keyword Search** (`mode="keyword"`): Uses keyword matching to find relevant chunks +3. **Hybrid Search** (`mode="hybrid"`): Combines both vector and keyword scores using a ranker + +### Hybrid Search + +Hybrid search combines the strengths of both vector and keyword search by: +- Computing vector similarity scores +- Computing keyword match scores +- Using a ranker to combine these scores + +Two ranker types are supported: + +1. **RRF (Reciprocal Rank Fusion)**: + - Combines ranks from both vector and keyword results + - Uses an impact factor (default: 60.0) to control the weight of higher-ranked results + - Good for balancing between vector and keyword results + - The default impact factor of 60.0 comes from the original RRF paper by Cormack et al. (2009) [^1], which found this value to provide optimal performance across various retrieval tasks + +2. **Weighted**: + - Linearly combines normalized vector and keyword scores + - Uses an alpha parameter (0-1) to control the blend: + - alpha=0: Only use keyword scores + - alpha=1: Only use vector scores + - alpha=0.5: Equal weight to both (default) + +Example using RAGQueryConfig with different search modes: + +```python +from llama_stack.apis.tools import RAGQueryConfig, RRFRanker, WeightedRanker + +# Vector search +config = RAGQueryConfig(mode="vector", max_chunks=5) + +# Keyword search +config = RAGQueryConfig(mode="keyword", max_chunks=5) + +# Hybrid search with custom RRF ranker +config = RAGQueryConfig( + mode="hybrid", + max_chunks=5, + ranker=RRFRanker(impact_factor=50.0), # Custom impact factor +) + +# Hybrid search with weighted ranker +config = RAGQueryConfig( + mode="hybrid", + max_chunks=5, + ranker=WeightedRanker(alpha=0.7), # 70% vector, 30% keyword +) + +# Hybrid search with default RRF ranker +config = RAGQueryConfig( + mode="hybrid", max_chunks=5 +) # Will use RRF with impact_factor=60.0 +``` + +Note: The ranker configuration is only used in hybrid mode. For vector or keyword modes, the ranker parameter is ignored. + +## Installation + +You can install SQLite-Vec using pip: + +```bash +pip install sqlite-vec +``` + +## Documentation + +See [sqlite-vec's GitHub repo](https://github.com/asg017/sqlite-vec/tree/main) for more details about sqlite-vec in general. + +[^1]: Cormack, G. V., Clarke, C. L., & Buettcher, S. (2009). [Reciprocal rank fusion outperforms condorcet and individual rank learning methods](https://dl.acm.org/doi/10.1145/1571941.1572114). In Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval (pp. 758-759). + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `db_path` | `` | No | | Path to the SQLite database file | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Config for KV store backend (SQLite only for now) | + +## Sample Configuration + +```yaml +db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/sqlite_vec.db +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/sqlite_vec_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/inline_sqlite_vec.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/inline_sqlite_vec.mdx new file mode 100644 index 0000000..7f69f61 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/inline_sqlite_vec.mdx @@ -0,0 +1,34 @@ +--- +description: "Please refer to the sqlite-vec provider documentation." +sidebar_label: Sqlite Vec +title: inline::sqlite_vec +--- + +# inline::sqlite_vec + +## Description + + +Please refer to the sqlite-vec provider documentation. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `db_path` | `` | No | | Path to the SQLite database file | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Config for KV store backend (SQLite only for now) | + +## Sample Configuration + +```yaml +db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/sqlite_vec.db +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/sqlite_vec_registry.db +``` +## Deprecation Notice + +:::warning +Please use the `inline::sqlite-vec` provider (notice the hyphen instead of underscore) instead. +::: diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/remote_chromadb.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/remote_chromadb.mdx new file mode 100644 index 0000000..8077710 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/remote_chromadb.mdx @@ -0,0 +1,90 @@ +--- +description: | + [Chroma](https://www.trychroma.com/) is an inline and remote vector + database provider for Llama Stack. It allows you to store and query vectors directly within a Chroma database. + That means you're not limited to storing vectors in memory or in a separate service. + + ## Features + Chroma supports: + - Store embeddings and their metadata + - Vector search + - Full-text search + - Document storage + - Metadata filtering + - Multi-modal retrieval + + ## Usage + + To use Chrome in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use chroma. + 3. Start storing and querying vectors. + + ## Installation + + You can install chroma using pip: + + ```bash + pip install chromadb + ``` + + ## Documentation + See [Chroma's documentation](https://docs.trychroma.com/docs/overview/introduction) for more details about Chroma in general. +sidebar_label: Remote - Chromadb +title: remote::chromadb +--- + +# remote::chromadb + +## Description + + +[Chroma](https://www.trychroma.com/) is an inline and remote vector +database provider for Llama Stack. It allows you to store and query vectors directly within a Chroma database. +That means you're not limited to storing vectors in memory or in a separate service. + +## Features +Chroma supports: +- Store embeddings and their metadata +- Vector search +- Full-text search +- Document storage +- Metadata filtering +- Multi-modal retrieval + +## Usage + +To use Chrome in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use chroma. +3. Start storing and querying vectors. + +## Installation + +You can install chroma using pip: + +```bash +pip install chromadb +``` + +## Documentation +See [Chroma's documentation](https://docs.trychroma.com/docs/overview/introduction) for more details about Chroma in general. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `url` | `str \| None` | No | | | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Config for KV store backend | + +## Sample Configuration + +```yaml +url: ${env.CHROMADB_URL} +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/chroma_remote_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/remote_milvus.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/remote_milvus.mdx new file mode 100644 index 0000000..7f7c081 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/remote_milvus.mdx @@ -0,0 +1,426 @@ +--- +description: | + [Milvus](https://milvus.io/) is an inline and remote vector database provider for Llama Stack. It + allows you to store and query vectors directly within a Milvus database. + That means you're not limited to storing vectors in memory or in a separate service. + + ## Features + + - Easy to use + - Fully integrated with Llama Stack + - Supports all search modes: vector, keyword, and hybrid search (both inline and remote configurations) + + ## Usage + + To use Milvus in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use Milvus. + 3. Start storing and querying vectors. + + ## Installation + + If you want to use inline Milvus, you can install: + + ```bash + pip install pymilvus[milvus-lite] + ``` + + If you want to use remote Milvus, you can install: + + ```bash + pip install pymilvus + ``` + + ## Configuration + + In Llama Stack, Milvus can be configured in two ways: + - **Inline (Local) Configuration** - Uses Milvus-Lite for local storage + - **Remote Configuration** - Connects to a remote Milvus server + + ### Inline (Local) Configuration + + The simplest method is local configuration, which requires setting `db_path`, a path for locally storing Milvus-Lite files: + + ```yaml + vector_io: + - provider_id: milvus + provider_type: inline::milvus + config: + db_path: ~/.llama/distributions/together/milvus_store.db + ``` + + ### Remote Configuration + + Remote configuration is suitable for larger data storage requirements: + + #### Standard Remote Connection + + ```yaml + vector_io: + - provider_id: milvus + provider_type: remote::milvus + config: + uri: "http://:" + token: ":" + ``` + + #### TLS-Enabled Remote Connection (One-way TLS) + + For connections to Milvus instances with one-way TLS enabled: + + ```yaml + vector_io: + - provider_id: milvus + provider_type: remote::milvus + config: + uri: "https://:" + token: ":" + secure: True + server_pem_path: "/path/to/server.pem" + ``` + + #### Mutual TLS (mTLS) Remote Connection + + For connections to Milvus instances with mutual TLS (mTLS) enabled: + + ```yaml + vector_io: + - provider_id: milvus + provider_type: remote::milvus + config: + uri: "https://:" + token: ":" + secure: True + ca_pem_path: "/path/to/ca.pem" + client_pem_path: "/path/to/client.pem" + client_key_path: "/path/to/client.key" + ``` + + #### Key Parameters for TLS Configuration + + - **`secure`**: Enables TLS encryption when set to `true`. Defaults to `false`. + - **`server_pem_path`**: Path to the **server certificate** for verifying the server's identity (used in one-way TLS). + - **`ca_pem_path`**: Path to the **Certificate Authority (CA) certificate** for validating the server certificate (required in mTLS). + - **`client_pem_path`**: Path to the **client certificate** file (required for mTLS). + - **`client_key_path`**: Path to the **client private key** file (required for mTLS). + + ## Search Modes + + Milvus supports three different search modes for both inline and remote configurations: + + ### Vector Search + Vector search uses semantic similarity to find the most relevant chunks based on embedding vectors. This is the default search mode and works well for finding conceptually similar content. + + ```python + # Vector search example + search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="What is machine learning?", + search_mode="vector", + max_num_results=5, + ) + ``` + + ### Keyword Search + Keyword search uses traditional text-based matching to find chunks containing specific terms or phrases. This is useful when you need exact term matches. + + ```python + # Keyword search example + search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="Python programming language", + search_mode="keyword", + max_num_results=5, + ) + ``` + + ### Hybrid Search + Hybrid search combines both vector and keyword search methods to provide more comprehensive results. It leverages the strengths of both semantic similarity and exact term matching. + + #### Basic Hybrid Search + ```python + # Basic hybrid search example (uses RRF ranker with default impact_factor=60.0) + search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="neural networks in Python", + search_mode="hybrid", + max_num_results=5, + ) + ``` + + **Note**: The default `impact_factor` value of 60.0 was empirically determined to be optimal in the original RRF research paper: ["Reciprocal Rank Fusion outperforms Condorcet and individual Rank Learning Methods"](https://plg.uwaterloo.ca/~gvcormac/cormacksigir09-rrf.pdf) (Cormack et al., 2009). + + #### Hybrid Search with RRF (Reciprocal Rank Fusion) Ranker + RRF combines rankings from vector and keyword search by using reciprocal ranks. The impact factor controls how much weight is given to higher-ranked results. + + ```python + # Hybrid search with custom RRF parameters + search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="neural networks in Python", + search_mode="hybrid", + max_num_results=5, + ranking_options={ + "ranker": { + "type": "rrf", + "impact_factor": 100.0, # Higher values give more weight to top-ranked results + } + }, + ) + ``` + + #### Hybrid Search with Weighted Ranker + Weighted ranker linearly combines normalized scores from vector and keyword search. The alpha parameter controls the balance between the two search methods. + + ```python + # Hybrid search with weighted ranker + search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="neural networks in Python", + search_mode="hybrid", + max_num_results=5, + ranking_options={ + "ranker": { + "type": "weighted", + "alpha": 0.7, # 70% vector search, 30% keyword search + } + }, + ) + ``` + + For detailed documentation on RRF and Weighted rankers, please refer to the [Milvus Reranking Guide](https://milvus.io/docs/reranking.md). + + ## Documentation + See the [Milvus documentation](https://milvus.io/docs/install-overview.md) for more details about Milvus in general. + + For more details on TLS configuration, refer to the [TLS setup guide](https://milvus.io/docs/tls.md). +sidebar_label: Remote - Milvus +title: remote::milvus +--- + +# remote::milvus + +## Description + + +[Milvus](https://milvus.io/) is an inline and remote vector database provider for Llama Stack. It +allows you to store and query vectors directly within a Milvus database. +That means you're not limited to storing vectors in memory or in a separate service. + +## Features + +- Easy to use +- Fully integrated with Llama Stack +- Supports all search modes: vector, keyword, and hybrid search (both inline and remote configurations) + +## Usage + +To use Milvus in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use Milvus. +3. Start storing and querying vectors. + +## Installation + +If you want to use inline Milvus, you can install: + +```bash +pip install pymilvus[milvus-lite] +``` + +If you want to use remote Milvus, you can install: + +```bash +pip install pymilvus +``` + +## Configuration + +In Llama Stack, Milvus can be configured in two ways: +- **Inline (Local) Configuration** - Uses Milvus-Lite for local storage +- **Remote Configuration** - Connects to a remote Milvus server + +### Inline (Local) Configuration + +The simplest method is local configuration, which requires setting `db_path`, a path for locally storing Milvus-Lite files: + +```yaml +vector_io: + - provider_id: milvus + provider_type: inline::milvus + config: + db_path: ~/.llama/distributions/together/milvus_store.db +``` + +### Remote Configuration + +Remote configuration is suitable for larger data storage requirements: + +#### Standard Remote Connection + +```yaml +vector_io: + - provider_id: milvus + provider_type: remote::milvus + config: + uri: "http://:" + token: ":" +``` + +#### TLS-Enabled Remote Connection (One-way TLS) + +For connections to Milvus instances with one-way TLS enabled: + +```yaml +vector_io: + - provider_id: milvus + provider_type: remote::milvus + config: + uri: "https://:" + token: ":" + secure: True + server_pem_path: "/path/to/server.pem" +``` + +#### Mutual TLS (mTLS) Remote Connection + +For connections to Milvus instances with mutual TLS (mTLS) enabled: + +```yaml +vector_io: + - provider_id: milvus + provider_type: remote::milvus + config: + uri: "https://:" + token: ":" + secure: True + ca_pem_path: "/path/to/ca.pem" + client_pem_path: "/path/to/client.pem" + client_key_path: "/path/to/client.key" +``` + +#### Key Parameters for TLS Configuration + +- **`secure`**: Enables TLS encryption when set to `true`. Defaults to `false`. +- **`server_pem_path`**: Path to the **server certificate** for verifying the server's identity (used in one-way TLS). +- **`ca_pem_path`**: Path to the **Certificate Authority (CA) certificate** for validating the server certificate (required in mTLS). +- **`client_pem_path`**: Path to the **client certificate** file (required for mTLS). +- **`client_key_path`**: Path to the **client private key** file (required for mTLS). + +## Search Modes + +Milvus supports three different search modes for both inline and remote configurations: + +### Vector Search +Vector search uses semantic similarity to find the most relevant chunks based on embedding vectors. This is the default search mode and works well for finding conceptually similar content. + +```python +# Vector search example +search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="What is machine learning?", + search_mode="vector", + max_num_results=5, +) +``` + +### Keyword Search +Keyword search uses traditional text-based matching to find chunks containing specific terms or phrases. This is useful when you need exact term matches. + +```python +# Keyword search example +search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="Python programming language", + search_mode="keyword", + max_num_results=5, +) +``` + +### Hybrid Search +Hybrid search combines both vector and keyword search methods to provide more comprehensive results. It leverages the strengths of both semantic similarity and exact term matching. + +#### Basic Hybrid Search +```python +# Basic hybrid search example (uses RRF ranker with default impact_factor=60.0) +search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="neural networks in Python", + search_mode="hybrid", + max_num_results=5, +) +``` + +**Note**: The default `impact_factor` value of 60.0 was empirically determined to be optimal in the original RRF research paper: ["Reciprocal Rank Fusion outperforms Condorcet and individual Rank Learning Methods"](https://plg.uwaterloo.ca/~gvcormac/cormacksigir09-rrf.pdf) (Cormack et al., 2009). + +#### Hybrid Search with RRF (Reciprocal Rank Fusion) Ranker +RRF combines rankings from vector and keyword search by using reciprocal ranks. The impact factor controls how much weight is given to higher-ranked results. + +```python +# Hybrid search with custom RRF parameters +search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="neural networks in Python", + search_mode="hybrid", + max_num_results=5, + ranking_options={ + "ranker": { + "type": "rrf", + "impact_factor": 100.0, # Higher values give more weight to top-ranked results + } + }, +) +``` + +#### Hybrid Search with Weighted Ranker +Weighted ranker linearly combines normalized scores from vector and keyword search. The alpha parameter controls the balance between the two search methods. + +```python +# Hybrid search with weighted ranker +search_response = client.vector_stores.search( + vector_store_id=vector_store.id, + query="neural networks in Python", + search_mode="hybrid", + max_num_results=5, + ranking_options={ + "ranker": { + "type": "weighted", + "alpha": 0.7, # 70% vector search, 30% keyword search + } + }, +) +``` + +For detailed documentation on RRF and Weighted rankers, please refer to the [Milvus Reranking Guide](https://milvus.io/docs/reranking.md). + +## Documentation +See the [Milvus documentation](https://milvus.io/docs/install-overview.md) for more details about Milvus in general. + +For more details on TLS configuration, refer to the [TLS setup guide](https://milvus.io/docs/tls.md). + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `uri` | `` | No | | The URI of the Milvus server | +| `token` | `str \| None` | No | | The token of the Milvus server | +| `consistency_level` | `` | No | Strong | The consistency level of the Milvus server | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | Config for KV store backend | +| `config` | `dict` | No | `{}` | This configuration allows additional fields to be passed through to the underlying Milvus client. See the [Milvus](https://milvus.io/docs/install-overview.md) documentation for more details about Milvus in general. | + +:::note +This configuration class accepts additional fields beyond those listed above. You can pass any additional configuration options that will be forwarded to the underlying provider. +::: + +## Sample Configuration + +```yaml +uri: ${env.MILVUS_ENDPOINT} +token: ${env.MILVUS_TOKEN} +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/milvus_remote_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/remote_pgvector.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/remote_pgvector.mdx new file mode 100644 index 0000000..d21810c --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/remote_pgvector.mdx @@ -0,0 +1,234 @@ +--- +description: | + [PGVector](https://github.com/pgvector/pgvector) is a remote vector database provider for Llama Stack. It + allows you to store and query vectors directly in memory. + That means you'll get fast and efficient vector retrieval. + + ## Features + + - Easy to use + - Fully integrated with Llama Stack + + There are three implementations of search for PGVectoIndex available: + + 1. Vector Search: + - How it works: + - Uses PostgreSQL's vector extension (pgvector) to perform similarity search + - Compares query embeddings against stored embeddings using Cosine distance or other distance metrics + - Eg. SQL query: SELECT document, embedding <=> %s::vector AS distance FROM table ORDER BY distance + + -Characteristics: + - Semantic understanding - finds documents similar in meaning even if they don't share keywords + - Works with high-dimensional vector embeddings (typically 768, 1024, or higher dimensions) + - Best for: Finding conceptually related content, handling synonyms, cross-language search + + 2. Keyword Search + - How it works: + - Uses PostgreSQL's full-text search capabilities with tsvector and ts_rank + - Converts text to searchable tokens using to_tsvector('english', text). Default language is English. + - Eg. SQL query: SELECT document, ts_rank(tokenized_content, plainto_tsquery('english', %s)) AS score + + - Characteristics: + - Lexical matching - finds exact keyword matches and variations + - Uses GIN (Generalized Inverted Index) for fast text search performance + - Scoring: Uses PostgreSQL's ts_rank function for relevance scoring + - Best for: Exact term matching, proper names, technical terms, Boolean-style queries + + 3. Hybrid Search + - How it works: + - Combines both vector and keyword search results + - Runs both searches independently, then merges results using configurable reranking + + - Two reranking strategies available: + - Reciprocal Rank Fusion (RRF) - (default: 60.0) + - Weighted Average - (default: 0.5) + + - Characteristics: + - Best of both worlds: semantic understanding + exact matching + - Documents appearing in both searches get boosted scores + - Configurable balance between semantic and lexical matching + - Best for: General-purpose search where you want both precision and recall + + 4. Database Schema + The PGVector implementation stores data optimized for all three search types: + CREATE TABLE vector_store_xxx ( + id TEXT PRIMARY KEY, + document JSONB, -- Original document + embedding vector(dimension), -- For vector search + content_text TEXT, -- Raw text content + tokenized_content TSVECTOR -- For keyword search + ); + + -- Indexes for performance + CREATE INDEX content_gin_idx ON table USING GIN(tokenized_content); -- Keyword search + -- Vector index created automatically by pgvector + + ## Usage + + To use PGVector in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use pgvector. (e.g. remote::pgvector). + 3. Start storing and querying vectors. + + ## This is an example how you can set up your environment for using PGVector + + 1. Export env vars: + ```bash + export ENABLE_PGVECTOR=true + export PGVECTOR_HOST=localhost + export PGVECTOR_PORT=5432 + export PGVECTOR_DB=llamastack + export PGVECTOR_USER=llamastack + export PGVECTOR_PASSWORD=llamastack + ``` + + 2. Create DB: + ```bash + psql -h localhost -U postgres -c "CREATE ROLE llamastack LOGIN PASSWORD 'llamastack';" + psql -h localhost -U postgres -c "CREATE DATABASE llamastack OWNER llamastack;" + psql -h localhost -U llamastack -d llamastack -c "CREATE EXTENSION IF NOT EXISTS vector;" + ``` + + ## Installation + + You can install PGVector using docker: + + ```bash + docker pull pgvector/pgvector:pg17 + ``` + ## Documentation + See [PGVector's documentation](https://github.com/pgvector/pgvector) for more details about PGVector in general. +sidebar_label: Remote - Pgvector +title: remote::pgvector +--- + +# remote::pgvector + +## Description + + +[PGVector](https://github.com/pgvector/pgvector) is a remote vector database provider for Llama Stack. It +allows you to store and query vectors directly in memory. +That means you'll get fast and efficient vector retrieval. + +## Features + +- Easy to use +- Fully integrated with Llama Stack + +There are three implementations of search for PGVectoIndex available: + +1. Vector Search: +- How it works: + - Uses PostgreSQL's vector extension (pgvector) to perform similarity search + - Compares query embeddings against stored embeddings using Cosine distance or other distance metrics + - Eg. SQL query: SELECT document, embedding <=> %s::vector AS distance FROM table ORDER BY distance + +-Characteristics: + - Semantic understanding - finds documents similar in meaning even if they don't share keywords + - Works with high-dimensional vector embeddings (typically 768, 1024, or higher dimensions) + - Best for: Finding conceptually related content, handling synonyms, cross-language search + +2. Keyword Search +- How it works: + - Uses PostgreSQL's full-text search capabilities with tsvector and ts_rank + - Converts text to searchable tokens using to_tsvector('english', text). Default language is English. + - Eg. SQL query: SELECT document, ts_rank(tokenized_content, plainto_tsquery('english', %s)) AS score + +- Characteristics: + - Lexical matching - finds exact keyword matches and variations + - Uses GIN (Generalized Inverted Index) for fast text search performance + - Scoring: Uses PostgreSQL's ts_rank function for relevance scoring + - Best for: Exact term matching, proper names, technical terms, Boolean-style queries + +3. Hybrid Search +- How it works: + - Combines both vector and keyword search results + - Runs both searches independently, then merges results using configurable reranking + +- Two reranking strategies available: + - Reciprocal Rank Fusion (RRF) - (default: 60.0) + - Weighted Average - (default: 0.5) + +- Characteristics: + - Best of both worlds: semantic understanding + exact matching + - Documents appearing in both searches get boosted scores + - Configurable balance between semantic and lexical matching + - Best for: General-purpose search where you want both precision and recall + +4. Database Schema +The PGVector implementation stores data optimized for all three search types: +CREATE TABLE vector_store_xxx ( + id TEXT PRIMARY KEY, + document JSONB, -- Original document + embedding vector(dimension), -- For vector search + content_text TEXT, -- Raw text content + tokenized_content TSVECTOR -- For keyword search +); + +-- Indexes for performance +CREATE INDEX content_gin_idx ON table USING GIN(tokenized_content); -- Keyword search +-- Vector index created automatically by pgvector + +## Usage + +To use PGVector in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use pgvector. (e.g. remote::pgvector). +3. Start storing and querying vectors. + +## This is an example how you can set up your environment for using PGVector + +1. Export env vars: +```bash +export ENABLE_PGVECTOR=true +export PGVECTOR_HOST=localhost +export PGVECTOR_PORT=5432 +export PGVECTOR_DB=llamastack +export PGVECTOR_USER=llamastack +export PGVECTOR_PASSWORD=llamastack +``` + +2. Create DB: +```bash +psql -h localhost -U postgres -c "CREATE ROLE llamastack LOGIN PASSWORD 'llamastack';" +psql -h localhost -U postgres -c "CREATE DATABASE llamastack OWNER llamastack;" +psql -h localhost -U llamastack -d llamastack -c "CREATE EXTENSION IF NOT EXISTS vector;" +``` + +## Installation + +You can install PGVector using docker: + +```bash +docker pull pgvector/pgvector:pg17 +``` +## Documentation +See [PGVector's documentation](https://github.com/pgvector/pgvector) for more details about PGVector in general. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `host` | `str \| None` | No | localhost | | +| `port` | `int \| None` | No | 5432 | | +| `db` | `str \| None` | No | postgres | | +| `user` | `str \| None` | No | postgres | | +| `password` | `str \| None` | No | mysecretpassword | | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig, annotation=NoneType, required=False, default='sqlite', discriminator='type'` | No | | Config for KV store backend (SQLite only for now) | + +## Sample Configuration + +```yaml +host: ${env.PGVECTOR_HOST:=localhost} +port: ${env.PGVECTOR_PORT:=5432} +db: ${env.PGVECTOR_DB} +user: ${env.PGVECTOR_USER} +password: ${env.PGVECTOR_PASSWORD} +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/pgvector_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/remote_qdrant.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/remote_qdrant.mdx new file mode 100644 index 0000000..c44a2b9 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/remote_qdrant.mdx @@ -0,0 +1,38 @@ +--- +description: "Please refer to the inline provider documentation." +sidebar_label: Remote - Qdrant +title: remote::qdrant +--- + +# remote::qdrant + +## Description + + +Please refer to the inline provider documentation. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `location` | `str \| None` | No | | | +| `url` | `str \| None` | No | | | +| `port` | `int \| None` | No | 6333 | | +| `grpc_port` | `` | No | 6334 | | +| `prefer_grpc` | `` | No | False | | +| `https` | `bool \| None` | No | | | +| `api_key` | `str \| None` | No | | | +| `prefix` | `str \| None` | No | | | +| `timeout` | `int \| None` | No | | | +| `host` | `str \| None` | No | | | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig` | No | sqlite | | + +## Sample Configuration + +```yaml +api_key: ${env.QDRANT_API_KEY:=} +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/qdrant_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/providers/vector_io/remote_weaviate.mdx b/versioned_docs/version-v0.2.23/providers/vector_io/remote_weaviate.mdx new file mode 100644 index 0000000..3f1e364 --- /dev/null +++ b/versioned_docs/version-v0.2.23/providers/vector_io/remote_weaviate.mdx @@ -0,0 +1,88 @@ +--- +description: | + [Weaviate](https://weaviate.io/) is a vector database provider for Llama Stack. + It allows you to store and query vectors directly within a Weaviate database. + That means you're not limited to storing vectors in memory or in a separate service. + + ## Features + Weaviate supports: + - Store embeddings and their metadata + - Vector search + - Full-text search + - Hybrid search + - Document storage + - Metadata filtering + - Multi-modal retrieval + + + ## Usage + + To use Weaviate in your Llama Stack project, follow these steps: + + 1. Install the necessary dependencies. + 2. Configure your Llama Stack project to use chroma. + 3. Start storing and querying vectors. + + ## Installation + + To install Weaviate see the [Weaviate quickstart documentation](https://weaviate.io/developers/weaviate/quickstart). + + ## Documentation + See [Weaviate's documentation](https://weaviate.io/developers/weaviate) for more details about Weaviate in general. +sidebar_label: Remote - Weaviate +title: remote::weaviate +--- + +# remote::weaviate + +## Description + + +[Weaviate](https://weaviate.io/) is a vector database provider for Llama Stack. +It allows you to store and query vectors directly within a Weaviate database. +That means you're not limited to storing vectors in memory or in a separate service. + +## Features +Weaviate supports: +- Store embeddings and their metadata +- Vector search +- Full-text search +- Hybrid search +- Document storage +- Metadata filtering +- Multi-modal retrieval + + +## Usage + +To use Weaviate in your Llama Stack project, follow these steps: + +1. Install the necessary dependencies. +2. Configure your Llama Stack project to use chroma. +3. Start storing and querying vectors. + +## Installation + +To install Weaviate see the [Weaviate quickstart documentation](https://weaviate.io/developers/weaviate/quickstart). + +## Documentation +See [Weaviate's documentation](https://weaviate.io/developers/weaviate) for more details about Weaviate in general. + + +## Configuration + +| Field | Type | Required | Default | Description | +|-------|------|----------|---------|-------------| +| `weaviate_api_key` | `str \| None` | No | | The API key for the Weaviate instance | +| `weaviate_cluster_url` | `str \| None` | No | localhost:8080 | The URL of the Weaviate cluster | +| `kvstore` | `utils.kvstore.config.RedisKVStoreConfig \| utils.kvstore.config.SqliteKVStoreConfig \| utils.kvstore.config.PostgresKVStoreConfig \| utils.kvstore.config.MongoDBKVStoreConfig, annotation=NoneType, required=False, default='sqlite', discriminator='type'` | No | | Config for KV store backend (SQLite only for now) | + +## Sample Configuration + +```yaml +weaviate_api_key: null +weaviate_cluster_url: ${env.WEAVIATE_CLUSTER_URL:=localhost:8080} +kvstore: + type: sqlite + db_path: ${env.SQLITE_STORE_DIR:=~/.llama/dummy}/weaviate_registry.db +``` diff --git a/versioned_docs/version-v0.2.23/references/evals_reference/index.mdx b/versioned_docs/version-v0.2.23/references/evals_reference/index.mdx new file mode 100644 index 0000000..0ec555e --- /dev/null +++ b/versioned_docs/version-v0.2.23/references/evals_reference/index.mdx @@ -0,0 +1,377 @@ +# Evaluations + +The Llama Stack Evaluation flow allows you to run evaluations on your GenAI application datasets or pre-registered benchmarks. + +We introduce a set of APIs in Llama Stack for supporting running evaluations of LLM applications. +- `/datasetio` + `/datasets` API +- `/scoring` + `/scoring_functions` API +- `/eval` + `/benchmarks` API + +This guide goes over the sets of APIs and developer experience flow of using Llama Stack to run evaluations for different use cases. Checkout our Colab notebook on working examples with evaluations [here](https://colab.research.google.com/drive/10CHyykee9j2OigaIcRv47BKG9mrNm0tJ?usp=sharing). + +## Evaluation Concepts + +The Evaluation APIs are associated with a set of Resources as shown in the following diagram. Please visit the Resources section in our [Core Concepts](../concepts/) guide for better high-level understanding. + +![Eval Concepts](/img/eval-concept.png) + +- **DatasetIO**: defines interface with datasets and data loaders. + - Associated with `Dataset` resource. +- **Scoring**: evaluate outputs of the system. + - Associated with `ScoringFunction` resource. We provide a suite of out-of-the box scoring functions and also the ability for you to add custom evaluators. These scoring functions are the core part of defining an evaluation task to output evaluation metrics. +- **Eval**: generate outputs (via Inference or Agents) and perform scoring. + - Associated with `Benchmark` resource. + +## Evaluation Examples Walkthrough + +[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb) + +It is best to open this notebook in Colab to follow along with the examples. + +### 1. Open Benchmark Model Evaluation + +This first example walks you through how to evaluate a model candidate served by Llama Stack on open benchmarks. We will use the following benchmark: +- [MMMU](https://arxiv.org/abs/2311.16502) (A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI)]: Benchmark designed to evaluate multimodal models. +- [SimpleQA](https://openai.com/index/introducing-simpleqa/): Benchmark designed to access models to answer short, fact-seeking questions. + +#### 1.1 Running MMMU +- We will use a pre-processed MMMU dataset from [llamastack/mmmu](https://huggingface.co/datasets/llamastack/mmmu). The preprocessing code is shown in this [GitHub Gist](https://gist.github.com/yanxi0830/118e9c560227d27132a7fd10e2c92840). The dataset is obtained by transforming the original [MMMU/MMMU](https://huggingface.co/datasets/MMMU/MMMU) dataset into correct format by `inference/chat-completion` API. + +```python +import datasets + +ds = datasets.load_dataset(path="llamastack/mmmu", name="Agriculture", split="dev") +ds = ds.select_columns(["chat_completion_input", "input_query", "expected_answer"]) +eval_rows = ds.to_pandas().to_dict(orient="records") +``` + +- Next, we will run evaluation on an model candidate, we will need to: + - Define a system prompt + - Define an EvalCandidate + - Run evaluate on the dataset + +```python +from rich.pretty import pprint +from tqdm import tqdm + +SYSTEM_PROMPT_TEMPLATE = """ +You are an expert in {subject} whose job is to answer questions from the user using images. + +First, reason about the correct answer. + +Then write the answer in the following format where X is exactly one of A,B,C,D: + +Answer: X + +Make sure X is one of A,B,C,D. + +If you are uncertain of the correct answer, guess the most likely one. +""" + +system_message = { + "role": "system", + "content": SYSTEM_PROMPT_TEMPLATE.format(subject=subset), +} + +# register the evaluation benchmark task with the dataset and scoring function +client.benchmarks.register( + benchmark_id="meta-reference::mmmu", + dataset_id=f"mmmu-{subset}-{split}", + scoring_functions=["basic::regex_parser_multiple_choice_answer"], +) + +response = client.eval.evaluate_rows( + benchmark_id="meta-reference::mmmu", + input_rows=eval_rows, + scoring_functions=["basic::regex_parser_multiple_choice_answer"], + benchmark_config={ + "eval_candidate": { + "type": "model", + "model": "meta-llama/Llama-3.2-90B-Vision-Instruct", + "sampling_params": { + "strategy": { + "type": "top_p", + "temperature": 1.0, + "top_p": 0.95, + }, + "max_tokens": 4096, + "repeat_penalty": 1.0, + }, + "system_message": system_message, + }, + }, +) +pprint(response) +``` + +#### 1.2. Running SimpleQA +- We will use a pre-processed SimpleQA dataset from [llamastack/evals](https://huggingface.co/datasets/llamastack/evals/viewer/evals__simpleqa) which is obtained by transforming the input query into correct format accepted by `inference/chat-completion` API. +- Since we will be using this same dataset in our next example for Agentic evaluation, we will register it using the `/datasets` API, and interact with it through `/datasetio` API. + +```python +simpleqa_dataset_id = "huggingface::simpleqa" + +_ = client.datasets.register( + purpose="eval/messages-answer", + source={ + "type": "uri", + "uri": "huggingface://datasets/llamastack/simpleqa?split=train", + }, + dataset_id=simpleqa_dataset_id, +) + +eval_rows = client.datasets.iterrows( + dataset_id=simpleqa_dataset_id, + limit=5, +) +``` + +```python +client.benchmarks.register( + benchmark_id="meta-reference::simpleqa", + dataset_id=simpleqa_dataset_id, + scoring_functions=["llm-as-judge::405b-simpleqa"], +) + +response = client.eval.evaluate_rows( + benchmark_id="meta-reference::simpleqa", + input_rows=eval_rows.data, + scoring_functions=["llm-as-judge::405b-simpleqa"], + benchmark_config={ + "eval_candidate": { + "type": "model", + "model": "meta-llama/Llama-3.2-90B-Vision-Instruct", + "sampling_params": { + "strategy": { + "type": "greedy", + }, + "max_tokens": 4096, + "repeat_penalty": 1.0, + }, + }, + }, +) +pprint(response) +``` + +### 2. Agentic Evaluation +- In this example, we will demonstrate how to evaluate a agent candidate served by Llama Stack via `/agent` API. +- We will continue to use the SimpleQA dataset we used in previous example. +- Instead of running evaluation on model, we will run the evaluation on a Search Agent with access to search tool. We will define our agent evaluation candidate through `AgentConfig`. + +```python +agent_config = { + "model": "meta-llama/Llama-3.3-70B-Instruct", + "instructions": "You are a helpful assistant that have access to tool to search the web. ", + "sampling_params": { + "strategy": { + "type": "top_p", + "temperature": 0.5, + "top_p": 0.9, + } + }, + "toolgroups": [ + "builtin::websearch", + ], + "tool_choice": "auto", + "tool_prompt_format": "json", + "input_shields": [], + "output_shields": [], + "enable_session_persistence": False, +} + +response = client.eval.evaluate_rows( + benchmark_id="meta-reference::simpleqa", + input_rows=eval_rows.data, + scoring_functions=["llm-as-judge::405b-simpleqa"], + benchmark_config={ + "eval_candidate": { + "type": "agent", + "config": agent_config, + }, + }, +) +pprint(response) +``` + +### 3. Agentic Application Dataset Scoring +[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb) + +Llama Stack offers a library of scoring functions and the `/scoring` API, allowing you to run evaluations on your pre-annotated AI application datasets. + +In this example, we will work with an example RAG dataset you have built previously, label with an annotation, and use LLM-As-Judge with custom judge prompt for scoring. Please checkout our [Llama Stack Playground](../building_applications/playground) for an interactive interface to upload datasets and run scorings. + +```python +judge_model_id = "meta-llama/Llama-3.1-405B-Instruct-FP8" + +JUDGE_PROMPT = """ +Given a QUESTION and GENERATED_RESPONSE and EXPECTED_RESPONSE. + +Compare the factual content of the GENERATED_RESPONSE with the EXPECTED_RESPONSE. Ignore any differences in style, grammar, or punctuation. + The GENERATED_RESPONSE may either be a subset or superset of the EXPECTED_RESPONSE, or it may conflict with it. Determine which case applies. Answer the question by selecting one of the following options: + (A) The GENERATED_RESPONSE is a subset of the EXPECTED_RESPONSE and is fully consistent with it. + (B) The GENERATED_RESPONSE is a superset of the EXPECTED_RESPONSE and is fully consistent with it. + (C) The GENERATED_RESPONSE contains all the same details as the EXPECTED_RESPONSE. + (D) There is a disagreement between the GENERATED_RESPONSE and the EXPECTED_RESPONSE. + (E) The answers differ, but these differences don't matter from the perspective of factuality. + +Give your answer in the format "Answer: One of ABCDE, Explanation: ". + +Your actual task: + +QUESTION: {input_query} +GENERATED_RESPONSE: {generated_answer} +EXPECTED_RESPONSE: {expected_answer} +""" + +input_query = ( + "What are the top 5 topics that were explained? Only list succinct bullet points." +) +generated_answer = """ +Here are the top 5 topics that were explained in the documentation for Torchtune: + +* What is LoRA and how does it work? +* Fine-tuning with LoRA: memory savings and parameter-efficient finetuning +* Running a LoRA finetune with Torchtune: overview and recipe +* Experimenting with different LoRA configurations: rank, alpha, and attention modules +* LoRA finetuning +""" +expected_answer = """LoRA""" + +dataset_rows = [ + { + "input_query": input_query, + "generated_answer": generated_answer, + "expected_answer": expected_answer, + }, +] + +scoring_params = { + "llm-as-judge::base": { + "judge_model": judge_model_id, + "prompt_template": JUDGE_PROMPT, + "type": "llm_as_judge", + "judge_score_regexes": ["Answer: (A|B|C|D|E)"], + }, + "basic::subset_of": None, + "braintrust::factuality": None, +} + +response = client.scoring.score( + input_rows=dataset_rows, scoring_functions=scoring_params +) +``` + +## Running Evaluations via CLI +The following examples give the quick steps to start running evaluations using the llama-stack-client CLI. + +### Benchmark Evaluation CLI +There are 3 necessary input for running a benchmark eval +- `list of benchmark_ids`: The list of benchmark ids to run evaluation on +- `model-id`: The model id to evaluate on +- `output_dir`: Path to store the evaluate results + +```bash +llama-stack-client eval run-benchmark ... \ +--model_id \ +--output_dir \ +``` + +You can run +```bash +llama-stack-client eval run-benchmark help +``` +to see the description of all the flags to run benchmark eval + +In the output log, you can find the path to the file that has your evaluation results. Open that file and you can see your aggregate evaluation results over there. + +### Application Evaluation CLI +Usage: For running application evals, you will already have available datasets in hand from your application. You will need to specify: +- `scoring-fn-id`: List of ScoringFunction identifiers you wish to use to run on your application. +- `Dataset` used for evaluation: + - (1) `--dataset-path`: path to local file system containing datasets to run evaluation on + - (2) `--dataset-id`: pre-registered dataset in Llama Stack +- (Optional) `--scoring-params-config`: optionally parameterize scoring functions with custom params (e.g. `judge_prompt`, `judge_model`, `parsing_regexes`). + +```bash +llama-stack-client eval run_scoring ... +--dataset-path \ +--output-dir ./ +``` + +### Defining BenchmarkConfig +The `BenchmarkConfig` are user specified config to define: +1. `EvalCandidate` to run generation on: + - `ModelCandidate`: The model will be used for generation through LlamaStack /inference API. + - `AgentCandidate`: The agentic system specified by AgentConfig will be used for generation through LlamaStack /agents API. +2. Optionally scoring function params to allow customization of scoring function behaviour. This is useful to parameterize generic scoring functions such as LLMAsJudge with custom `judge_model` / `judge_prompt`. + +**Example BenchmarkConfig** +```json +{ + "eval_candidate": { + "type": "model", + "model": "Llama3.1-405B-Instruct", + "sampling_params": { + "strategy": { + "type": "greedy", + }, + "max_tokens": 0, + "repetition_penalty": 1.0 + } + }, + "scoring_params": { + "llm-as-judge::llm_as_judge_base": { + "type": "llm_as_judge", + "judge_model": "meta-llama/Llama-3.1-8B-Instruct", + "prompt_template": "Your job is to look at a question, a gold target ........", + "judge_score_regexes": [ + "(A|B|C)" + ] + } + } +} +``` + +## Open-benchmark Contributing Guide + +### Create the new dataset for your new benchmark +An eval open-benchmark essentially contains 2 parts: +- `raw data`: The raw dataset associated with the benchmark. You typically need to search the original paper that introduces the benchmark and find the canonical dataset (usually hosted on huggingface) +- `prompt template`: How to ask the candidate model to generate the answer (prompt template plays a critical role to the evaluation results). Typically, you can find the reference prompt template associated with the benchmark in benchmarks author's repo ([example](https://github.com/idavidrein/gpqa/blob/main/prompts/chain_of_thought.txt)) or some other popular open source repos ([example](https://github.com/openai/simple-evals/blob/0a6e8f62e52bc5ae915f752466be3af596caf392/common.py#L14)) + +To create new open-benchmark in llama stack, you need to combine the prompt template and the raw data into the `chat_completion_input` column in the evaluation dataset. + +Llama stack enforces the evaluate dataset schema to contain at least 3 columns: +- `chat_completion_input`: The actual input to the model to run the generation for eval +- `input_query`: The raw input from the raw dataset without the prompt template +- `expected_answer`: The ground truth for scoring functions to calculate the score from. + +You need to write a script [example convert script](https://gist.github.com/yanxi0830/118e9c560227d27132a7fd10e2c92840) to convert the benchmark raw dataset to llama stack format eval dataset and update the dataset to huggingface [example benchmark dataset](https://huggingface.co/datasets/llamastack/mmmu) + +### Find scoring function for your new benchmark +The purpose of scoring function is to calculate the score for each example based on candidate model generation result and expected_answer. It also aggregates the scores from all the examples and generate the final evaluate results. + +Firstly, you can see if the existing [llama stack scoring functions](https://github.com/meta-llama/llama-stack/tree/main/llama_stack/providers/inline/scoring) can fulfill your need. If not, you need to write a new scoring function based on what benchmark author / other open source repo describe. + +### Add new benchmark into template +Firstly, you need to add the evaluation dataset associated with your benchmark under `datasets` resource in the [open-benchmark](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/distributions/open-benchmark/run.yaml) + +Secondly, you need to add the new benchmark you just created under the `benchmarks` resource in the same template. To add the new benchmark, you need to have +- `benchmark_id`: identifier of the benchmark +- `dataset_id`: identifier of the dataset associated with your benchmark +- `scoring_functions`: scoring function to calculate the score based on generation results and expected_answer + +### Test the new benchmark + +Spin up llama stack server with 'open-benchmark' templates +```bash +llama stack run llama_stack/distributions/open-benchmark/run.yaml +``` + +Run eval benchmark CLI with your new benchmark id +```bash +llama-stack-client eval run-benchmark \ +--model_id \ +--output_dir \ +``` diff --git a/versioned_docs/version-v0.2.23/references/index.mdx b/versioned_docs/version-v0.2.23/references/index.mdx new file mode 100644 index 0000000..dd6ab21 --- /dev/null +++ b/versioned_docs/version-v0.2.23/references/index.mdx @@ -0,0 +1,12 @@ +--- +title: References +description: Reference documentation for Llama Stack +sidebar_label: Overview +sidebar_position: 1 +--- + +# References + +- [Python SDK Reference](/docs/references/python_sdk_reference/) +- [Llama CLI](/docs/references/llama_cli_reference/) for building and running your Llama Stack server +- [Llama Stack Client CLI](./llama_stack_client_cli_reference.md) for interacting with your Llama Stack server diff --git a/versioned_docs/version-v0.2.23/references/llama_cli_reference/download_models.md b/versioned_docs/version-v0.2.23/references/llama_cli_reference/download_models.md new file mode 100644 index 0000000..a9af653 --- /dev/null +++ b/versioned_docs/version-v0.2.23/references/llama_cli_reference/download_models.md @@ -0,0 +1,165 @@ +# Downloading Models + +The `llama` CLI tool helps you setup and use the Llama Stack. It should be available on your path after installing the `llama-stack` package. + +## Installation + +You have two ways to install Llama Stack: + +1. **Install as a package**: + You can install the repository directly from [PyPI](https://pypi.org/project/llama-stack/) by running the following command: + ```bash + pip install llama-stack + ``` + +2. **Install from source**: + If you prefer to install from the source code, follow these steps: + ```bash + mkdir -p ~/local + cd ~/local + git clone git@github.com:meta-llama/llama-stack.git + + uv venv myenv --python 3.12 + source myenv/bin/activate # On Windows: myenv\Scripts\activate + + cd llama-stack + pip install -e . + +## Downloading models via CLI + +You first need to have models downloaded locally. + +To download any model you need the **Model Descriptor**. +This can be obtained by running the command +``` +llama model list +``` + +You should see a table like this: + +``` ++----------------------------------+------------------------------------------+----------------+ +| Model Descriptor(ID) | Hugging Face Repo | Context Length | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-8B | meta-llama/Llama-3.1-8B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-70B | meta-llama/Llama-3.1-70B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B:bf16-mp8 | meta-llama/Llama-3.1-405B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B | meta-llama/Llama-3.1-405B-FP8 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B:bf16-mp16 | meta-llama/Llama-3.1-405B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-8B-Instruct | meta-llama/Llama-3.1-8B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-70B-Instruct | meta-llama/Llama-3.1-70B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B-Instruct:bf16-mp8 | meta-llama/Llama-3.1-405B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B-Instruct | meta-llama/Llama-3.1-405B-Instruct-FP8 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B-Instruct:bf16-mp16 | meta-llama/Llama-3.1-405B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-1B | meta-llama/Llama-3.2-1B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-3B | meta-llama/Llama-3.2-3B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-11B-Vision | meta-llama/Llama-3.2-11B-Vision | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-90B-Vision | meta-llama/Llama-3.2-90B-Vision | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-1B-Instruct | meta-llama/Llama-3.2-1B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-3B-Instruct | meta-llama/Llama-3.2-3B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-11B-Vision-Instruct | meta-llama/Llama-3.2-11B-Vision-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-90B-Vision-Instruct | meta-llama/Llama-3.2-90B-Vision-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-11B-Vision | meta-llama/Llama-Guard-3-11B-Vision | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-1B:int4-mp1 | meta-llama/Llama-Guard-3-1B-INT4 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-1B | meta-llama/Llama-Guard-3-1B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-8B | meta-llama/Llama-Guard-3-8B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-8B:int8-mp1 | meta-llama/Llama-Guard-3-8B-INT8 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Prompt-Guard-86M | meta-llama/Prompt-Guard-86M | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-2-8B | meta-llama/Llama-Guard-2-8B | 4K | ++----------------------------------+------------------------------------------+----------------+ +``` + +To download models, you can use the llama download command. + +#### Downloading from [Meta](https://llama.meta.com/llama-downloads/) + +Here is an example download command to get the 3B-Instruct/11B-Vision-Instruct model. You will need META_URL which can be obtained from [here](https://llama.meta.com/docs/getting_the_models/meta/). Note: You need to quote the META_URL + +Download the required checkpoints using the following commands: +```bash +# download the 8B model, this can be run on a single GPU +llama download --source meta --model-id Llama3.2-3B-Instruct --meta-url 'META_URL' + +# you can also get the 70B model, this will require 8 GPUs however +llama download --source meta --model-id Llama3.2-11B-Vision-Instruct --meta-url 'META_URL' + +# llama-agents have safety enabled by default. For this, you will need +# safety models -- Llama-Guard and Prompt-Guard +llama download --source meta --model-id Prompt-Guard-86M --meta-url 'META_URL' +llama download --source meta --model-id Llama-Guard-3-1B --meta-url 'META_URL' +``` + +#### Downloading from [Hugging Face](https://huggingface.co/meta-llama) + +Essentially, the same commands above work, just replace `--source meta` with `--source huggingface`. + +```bash +llama download --source huggingface --model-id Llama3.1-8B-Instruct --hf-token + +llama download --source huggingface --model-id Llama3.1-70B-Instruct --hf-token + +llama download --source huggingface --model-id Llama-Guard-3-1B --ignore-patterns *original* +llama download --source huggingface --model-id Prompt-Guard-86M --ignore-patterns *original* +``` + +**Important:** Set your environment variable `HF_TOKEN` or pass in `--hf-token` to the command to validate your access. You can find your token at [https://huggingface.co/settings/tokens](https://huggingface.co/settings/tokens). + +```{tip} +Default for `llama download` is to run with `--ignore-patterns *.safetensors` since we use the `.pth` files in the `original` folder. For Llama Guard and Prompt Guard, however, we need safetensors. Hence, please run with `--ignore-patterns original` so that safetensors are downloaded and `.pth` files are ignored. +``` + +## List the downloaded models + +To list the downloaded models with the following command: +``` +llama model list --downloaded +``` + +You should see a table like this: +``` +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ Model โ”ƒ Size โ”ƒ Modified Time โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ Llama3.2-1B-Instruct:int4-qlora-eo8 โ”‚ 1.53 GB โ”‚ 2025-02-26 11:22:28 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-1B โ”‚ 2.31 GB โ”‚ 2025-02-18 21:48:52 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Prompt-Guard-86M โ”‚ 0.02 GB โ”‚ 2025-02-26 11:29:28 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-3B-Instruct:int4-spinquant-eo8 โ”‚ 3.69 GB โ”‚ 2025-02-26 11:37:41 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-3B โ”‚ 5.99 GB โ”‚ 2025-02-18 21:51:26 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.1-8B โ”‚ 14.97 GB โ”‚ 2025-02-16 10:36:37 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-1B-Instruct:int4-spinquant-eo8 โ”‚ 1.51 GB โ”‚ 2025-02-26 11:35:02 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama-Guard-3-1B โ”‚ 2.80 GB โ”‚ 2025-02-26 11:20:46 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama-Guard-3-1B:int4 โ”‚ 0.43 GB โ”‚ 2025-02-26 11:33:33 โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ +``` diff --git a/versioned_docs/version-v0.2.23/references/llama_cli_reference/index.md b/versioned_docs/version-v0.2.23/references/llama_cli_reference/index.md new file mode 100644 index 0000000..fe3aa51 --- /dev/null +++ b/versioned_docs/version-v0.2.23/references/llama_cli_reference/index.md @@ -0,0 +1,276 @@ +# llama (server-side) CLI Reference + +The `llama` CLI tool helps you set up and use the Llama Stack. The CLI is available on your path after installing the `llama-stack` package. + +## Installation + +You have two ways to install Llama Stack: + +1. **Install as a package**: + You can install the repository directly from [PyPI](https://pypi.org/project/llama-stack/) by running the following command: + ```bash + pip install llama-stack + ``` + +2. **Install from source**: + If you prefer to install from the source code, follow these steps: + ```bash + mkdir -p ~/local + cd ~/local + git clone git@github.com:meta-llama/llama-stack.git + + uv venv myenv --python 3.12 + source myenv/bin/activate # On Windows: myenv\Scripts\activate + + cd llama-stack + pip install -e . + + +## `llama` subcommands +1. `download`: Supports downloading models from Meta or Hugging Face. [Downloading models](#downloading-models) +2. `model`: Lists available models and their properties. [Understanding models](#understand-the-models) +3. `stack`: Allows you to build a stack using the `llama stack` distribution and run a Llama Stack server. You can read more about how to build a Llama Stack distribution in the [Build your own Distribution](../distributions/building_distro) documentation. + +### Sample Usage + +``` +llama --help +``` + +``` +usage: llama [-h] {download,model,stack} ... + +Welcome to the Llama CLI + +options: + -h, --help show this help message and exit + +subcommands: + {download,model,stack} +``` + +## Downloading models + +You first need to have models downloaded locally. + +To download any model you need the **Model Descriptor**. +This can be obtained by running the command +``` +llama model list +``` + +You should see a table like this: + +``` ++----------------------------------+------------------------------------------+----------------+ +| Model Descriptor(ID) | Hugging Face Repo | Context Length | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-8B | meta-llama/Llama-3.1-8B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-70B | meta-llama/Llama-3.1-70B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B:bf16-mp8 | meta-llama/Llama-3.1-405B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B | meta-llama/Llama-3.1-405B-FP8 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B:bf16-mp16 | meta-llama/Llama-3.1-405B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-8B-Instruct | meta-llama/Llama-3.1-8B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-70B-Instruct | meta-llama/Llama-3.1-70B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B-Instruct:bf16-mp8 | meta-llama/Llama-3.1-405B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B-Instruct | meta-llama/Llama-3.1-405B-Instruct-FP8 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.1-405B-Instruct:bf16-mp16 | meta-llama/Llama-3.1-405B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-1B | meta-llama/Llama-3.2-1B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-3B | meta-llama/Llama-3.2-3B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-11B-Vision | meta-llama/Llama-3.2-11B-Vision | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-90B-Vision | meta-llama/Llama-3.2-90B-Vision | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-1B-Instruct | meta-llama/Llama-3.2-1B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-3B-Instruct | meta-llama/Llama-3.2-3B-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-11B-Vision-Instruct | meta-llama/Llama-3.2-11B-Vision-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama3.2-90B-Vision-Instruct | meta-llama/Llama-3.2-90B-Vision-Instruct | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-11B-Vision | meta-llama/Llama-Guard-3-11B-Vision | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-1B:int4-mp1 | meta-llama/Llama-Guard-3-1B-INT4 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-1B | meta-llama/Llama-Guard-3-1B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-8B | meta-llama/Llama-Guard-3-8B | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-3-8B:int8-mp1 | meta-llama/Llama-Guard-3-8B-INT8 | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Prompt-Guard-86M | meta-llama/Prompt-Guard-86M | 128K | ++----------------------------------+------------------------------------------+----------------+ +| Llama-Guard-2-8B | meta-llama/Llama-Guard-2-8B | 4K | ++----------------------------------+------------------------------------------+----------------+ +``` + +To download models, you can use the `llama download` command. + +### Downloading from [Meta](https://llama.meta.com/llama-downloads/) + +Here is an example download command to get the 3B-Instruct/11B-Vision-Instruct model. You will need META_URL which can be obtained from [here](https://llama.meta.com/docs/getting_the_models/meta/) + +Download the required checkpoints using the following commands: +```bash +# download the 8B model, this can be run on a single GPU +llama download --source meta --model-id Llama3.2-3B-Instruct --meta-url META_URL + +# you can also get the 70B model, this will require 8 GPUs however +llama download --source meta --model-id Llama3.2-11B-Vision-Instruct --meta-url META_URL + +# llama-agents have safety enabled by default. For this, you will need +# safety models -- Llama-Guard and Prompt-Guard +llama download --source meta --model-id Prompt-Guard-86M --meta-url META_URL +llama download --source meta --model-id Llama-Guard-3-1B --meta-url META_URL +``` + +### Downloading from [Hugging Face](https://huggingface.co/meta-llama) + +Essentially, the same commands above work, just replace `--source meta` with `--source huggingface`. + +```bash +llama download --source huggingface --model-id Llama3.1-8B-Instruct --hf-token + +llama download --source huggingface --model-id Llama3.1-70B-Instruct --hf-token + +llama download --source huggingface --model-id Llama-Guard-3-1B --ignore-patterns *original* +llama download --source huggingface --model-id Prompt-Guard-86M --ignore-patterns *original* +``` + +**Important:** Set your environment variable `HF_TOKEN` or pass in `--hf-token` to the command to validate your access. You can find your token at [https://huggingface.co/settings/tokens](https://huggingface.co/settings/tokens). + +```{tip} +Default for `llama download` is to run with `--ignore-patterns *.safetensors` since we use the `.pth` files in the `original` folder. For Llama Guard and Prompt Guard, however, we need safetensors. Hence, please run with `--ignore-patterns original` so that safetensors are downloaded and `.pth` files are ignored. +``` + +## List the downloaded models + +To list the downloaded models with the following command: +``` +llama model list --downloaded +``` + +You should see a table like this: +``` +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ Model โ”ƒ Size โ”ƒ Modified Time โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ Llama3.2-1B-Instruct:int4-qlora-eo8 โ”‚ 1.53 GB โ”‚ 2025-02-26 11:22:28 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-1B โ”‚ 2.31 GB โ”‚ 2025-02-18 21:48:52 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Prompt-Guard-86M โ”‚ 0.02 GB โ”‚ 2025-02-26 11:29:28 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-3B-Instruct:int4-spinquant-eo8 โ”‚ 3.69 GB โ”‚ 2025-02-26 11:37:41 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-3B โ”‚ 5.99 GB โ”‚ 2025-02-18 21:51:26 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.1-8B โ”‚ 14.97 GB โ”‚ 2025-02-16 10:36:37 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama3.2-1B-Instruct:int4-spinquant-eo8 โ”‚ 1.51 GB โ”‚ 2025-02-26 11:35:02 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama-Guard-3-1B โ”‚ 2.80 GB โ”‚ 2025-02-26 11:20:46 โ”‚ +โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค +โ”‚ Llama-Guard-3-1B:int4 โ”‚ 0.43 GB โ”‚ 2025-02-26 11:33:33 โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ +``` + + +## Understand the models +The `llama model` command helps you explore the modelโ€™s interface. + +1. `download`: Download the model from different sources. (meta, huggingface) +2. `list`: Lists all the models available for download with hardware requirements for deploying the models. +3. `prompt-format`: Show llama model message formats. +4. `describe`: Describes all the properties of the model. + +### Sample Usage + +`llama model ` + +``` +llama model --help +``` +``` +usage: llama model [-h] {download,list,prompt-format,describe,verify-download,remove} ... + +Work with llama models + +options: + -h, --help show this help message and exit + +model_subcommands: + {download,list,prompt-format,describe,verify-download,remove} +``` + +### Describe + +You can use the describe command to know more about a model: +``` +llama model describe -m Llama3.2-3B-Instruct +``` +``` ++-----------------------------+----------------------------------+ +| Model | Llama3.2-3B-Instruct | ++-----------------------------+----------------------------------+ +| Hugging Face ID | meta-llama/Llama-3.2-3B-Instruct | ++-----------------------------+----------------------------------+ +| Description | Llama 3.2 3b instruct model | ++-----------------------------+----------------------------------+ +| Context Length | 128K tokens | ++-----------------------------+----------------------------------+ +| Weights format | bf16 | ++-----------------------------+----------------------------------+ +| Model params.json | { | +| | "dim": 3072, | +| | "n_layers": 28, | +| | "n_heads": 24, | +| | "n_kv_heads": 8, | +| | "vocab_size": 128256, | +| | "ffn_dim_multiplier": 1.0, | +| | "multiple_of": 256, | +| | "norm_eps": 1e-05, | +| | "rope_theta": 500000.0, | +| | "use_scaled_rope": true | +| | } | ++-----------------------------+----------------------------------+ +| Recommended sampling params | { | +| | "temperature": 1.0, | +| | "top_p": 0.9, | +| | "top_k": 0 | +| | } | ++-----------------------------+----------------------------------+ +``` + +### Prompt Format +You can even run `llama model prompt-format` see all of the templates and their tokens: + +``` +llama model prompt-format -m Llama3.2-3B-Instruct +``` +![alt text](../../../resources/prompt-format.png) + + +You will be shown a Markdown formatted description of the model interface and how prompts / messages are formatted for various scenarios. + +**NOTE**: Outputs in terminal are color printed to show special tokens. + +### Remove model +You can run `llama model remove` to remove an unnecessary model: + +``` +llama model remove -m Llama-Guard-3-8B-int8 +``` diff --git a/versioned_docs/version-v0.2.23/references/llama_stack_client_cli_reference.md b/versioned_docs/version-v0.2.23/references/llama_stack_client_cli_reference.md new file mode 100644 index 0000000..d4d79ce --- /dev/null +++ b/versioned_docs/version-v0.2.23/references/llama_stack_client_cli_reference.md @@ -0,0 +1,589 @@ +# llama (client-side) CLI Reference + +The `llama-stack-client` CLI allows you to query information about the distribution. + +## Basic Commands + +### `llama-stack-client` +```bash +llama-stack-client +Usage: llama-stack-client [OPTIONS] COMMAND [ARGS]... + + Welcome to the llama-stack-client CLI - a command-line interface for + interacting with Llama Stack + +Options: + --version Show the version and exit. + --endpoint TEXT Llama Stack distribution endpoint + --api-key TEXT Llama Stack distribution API key + --config TEXT Path to config file + --help Show this message and exit. + +Commands: + configure Configure Llama Stack Client CLI. + datasets Manage datasets. + eval Run evaluation tasks. + eval_tasks Manage evaluation tasks. + inference Inference (chat). + inspect Inspect server configuration. + models Manage GenAI models. + post_training Post-training. + providers Manage API providers. + scoring_functions Manage scoring functions. + shields Manage safety shield services. + toolgroups Manage available tool groups. + vector_dbs Manage vector databases. +``` + +### `llama-stack-client configure` +Configure Llama Stack Client CLI. +```bash +llama-stack-client configure +> Enter the host name of the Llama Stack distribution server: localhost +> Enter the port number of the Llama Stack distribution server: 8321 +Done! You can now use the Llama Stack Client CLI with endpoint http://localhost:8321 +``` + +Optional arguments: +- `--endpoint`: Llama Stack distribution endpoint +- `--api-key`: Llama Stack distribution API key + + + +## `llama-stack-client inspect version` +Inspect server configuration. +```bash +llama-stack-client inspect version +``` +```bash +VersionInfo(version='0.2.14') +``` + + +### `llama-stack-client providers list` +Show available providers on distribution endpoint +```bash +llama-stack-client providers list +``` +``` ++-----------+----------------+-----------------+ +| API | Provider ID | Provider Type | ++===========+================+=================+ +| scoring | meta0 | meta-reference | ++-----------+----------------+-----------------+ +| datasetio | meta0 | meta-reference | ++-----------+----------------+-----------------+ +| inference | tgi0 | remote::tgi | ++-----------+----------------+-----------------+ +| memory | meta-reference | meta-reference | ++-----------+----------------+-----------------+ +| agents | meta-reference | meta-reference | ++-----------+----------------+-----------------+ +| telemetry | meta-reference | meta-reference | ++-----------+----------------+-----------------+ +| safety | meta-reference | meta-reference | ++-----------+----------------+-----------------+ +``` + +### `llama-stack-client providers inspect` +Show specific provider configuration on distribution endpoint +```bash +llama-stack-client providers inspect +``` + + +## Inference +Inference (chat). + + +### `llama-stack-client inference chat-completion` +Show available inference chat completion endpoints on distribution endpoint +```bash +llama-stack-client inference chat-completion --message [--stream] [--session] [--model-id] +``` +```bash +OpenAIChatCompletion( + id='chatcmpl-aacd11f3-8899-4ec5-ac5b-e655132f6891', + choices=[ + OpenAIChatCompletionChoice( + finish_reason='stop', + index=0, + message=OpenAIChatCompletionChoiceMessageOpenAIAssistantMessageParam( + role='assistant', + content='The captain of the whaleship Pequod in Nathaniel Hawthorne\'s novel "Moby-Dick" is Captain +Ahab. He\'s a vengeful and obsessive old sailor who\'s determined to hunt down and kill the white sperm whale +Moby-Dick, whom he\'s lost his leg to in a previous encounter.', + name=None, + tool_calls=None, + refusal=None, + annotations=None, + audio=None, + function_call=None + ), + logprobs=None + ) + ], + created=1752578797, + model='llama3.2:3b-instruct-fp16', + object='chat.completion', + service_tier=None, + system_fingerprint='fp_ollama', + usage={ + 'completion_tokens': 67, + 'prompt_tokens': 33, + 'total_tokens': 100, + 'completion_tokens_details': None, + 'prompt_tokens_details': None + } +) +``` + +Required arguments: +**Note:** At least one of these parameters is required for chat completion +- `--message`: Message +- `--session`: Start a Chat Session + +Optional arguments: +- `--stream`: Stream +- `--model-id`: Model ID + +## Model Management +Manage GenAI models. + + +### `llama-stack-client models list` +Show available llama models at distribution endpoint +```bash +llama-stack-client models list +``` +``` +Available Models + +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ model_type โ”ƒ identifier โ”ƒ provider_resource_id โ”ƒ metadata โ”ƒ provider_id โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ llm โ”‚ meta-llama/Llama-3.2-3B-Instruct โ”‚ llama3.2:3b-instruct-fp16 โ”‚ โ”‚ ollama โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ + +Total models: 1 +``` + +### `llama-stack-client models get` +Show details of a specific model at the distribution endpoint +```bash +llama-stack-client models get Llama3.1-8B-Instruct +``` + +``` ++----------------------+----------------------+----------------------------------------------------------+---------------+ +| identifier | llama_model | metadata | provider_id | ++======================+======================+==========================================================+===============+ +| Llama3.1-8B-Instruct | Llama3.1-8B-Instruct | {'huggingface_repo': 'meta-llama/Llama-3.1-8B-Instruct'} | tgi0 | ++----------------------+----------------------+----------------------------------------------------------+---------------+ +``` + + +```bash +llama-stack-client models get Random-Model + +Model RandomModel is not found at distribution endpoint host:port. Please ensure endpoint is serving specified model. +``` + +### `llama-stack-client models register` +Register a new model at distribution endpoint +```bash +llama-stack-client models register [--provider-id ] [--provider-model-id ] [--metadata ] [--model-type ] +``` + +Required arguments: +- `MODEL_ID`: Model ID +- `--provider-id`: Provider ID for the model + +Optional arguments: +- `--provider-model-id`: Provider's model ID +- `--metadata`: JSON metadata for the model +- `--model-type`: Model type: `llm`, `embedding` + + +### `llama-stack-client models unregister` +Unregister a model from distribution endpoint +```bash +llama-stack-client models unregister +``` + +## Vector DB Management +Manage vector databases. + + +### `llama-stack-client vector_dbs list` +Show available vector dbs on distribution endpoint +```bash +llama-stack-client vector_dbs list +``` +``` +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ identifier โ”ƒ provider_id โ”ƒ provider_resource_id โ”ƒ vector_db_type โ”ƒ params โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ my_demo_vector_db โ”‚ faiss โ”‚ my_demo_vector_db โ”‚ โ”‚ embedding_dimension: 384 โ”‚ +โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ embedding_model: all-MiniLM-L6-v2 โ”‚ +โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ type: vector_db โ”‚ +โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ +``` + +### `llama-stack-client vector_dbs register` +Create a new vector db +```bash +llama-stack-client vector_dbs register [--provider-id ] [--provider-vector-db-id ] [--embedding-model ] [--embedding-dimension ] +``` + + +Required arguments: +- `VECTOR_DB_ID`: Vector DB ID + +Optional arguments: +- `--provider-id`: Provider ID for the vector db +- `--provider-vector-db-id`: Provider's vector db ID +- `--embedding-model`: Embedding model to use. Default: `all-MiniLM-L6-v2` +- `--embedding-dimension`: Dimension of embeddings. Default: 384 + +### `llama-stack-client vector_dbs unregister` +Delete a vector db +```bash +llama-stack-client vector_dbs unregister +``` + + +Required arguments: +- `VECTOR_DB_ID`: Vector DB ID + + +## Shield Management +Manage safety shield services. +### `llama-stack-client shields list` +Show available safety shields on distribution endpoint +```bash +llama-stack-client shields list +``` + +``` +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ identifier โ”ƒ provider_alias โ”ƒ params โ”ƒ provider_id โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ ollama โ”‚ ollama/llama-guard3:1b โ”‚ โ”‚ llama-guard โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ +``` + +### `llama-stack-client shields register` +Register a new safety shield +```bash +llama-stack-client shields register --shield-id [--provider-id ] [--provider-shield-id ] [--params ] +``` + +Required arguments: +- `--shield-id`: ID of the shield + +Optional arguments: +- `--provider-id`: Provider ID for the shield +- `--provider-shield-id`: Provider's shield ID +- `--params`: JSON configuration parameters for the shield + + +## Eval execution +Run evaluation tasks. + + +### `llama-stack-client eval run-benchmark` +Run a evaluation benchmark task +```bash +llama-stack-client eval run-benchmark [ ...] --eval-task-config --output-dir --model-id [--num-examples ] [--visualize] [--repeat-penalty ] [--top-p ] [--max-tokens ] +``` + +Required arguments: +- `--eval-task-config`: Path to the eval task config file in JSON format +- `--output-dir`: Path to the directory where evaluation results will be saved +- `--model-id`: model id to run the benchmark eval on + +Optional arguments: +- `--num-examples`: Number of examples to evaluate (useful for debugging) +- `--visualize`: If set, visualizes evaluation results after completion +- `--repeat-penalty`: repeat-penalty in the sampling params to run generation +- `--top-p`: top-p in the sampling params to run generation +- `--max-tokens`: max-tokens in the sampling params to run generation +- `--temperature`: temperature in the sampling params to run generation + +Example benchmark_config.json: +```json +{ + "type": "benchmark", + "eval_candidate": { + "type": "model", + "model": "Llama3.1-405B-Instruct", + "sampling_params": { + "strategy": "greedy", + } + } +} +``` + +### `llama-stack-client eval run-scoring` +Run scoring from application datasets +```bash +llama-stack-client eval run-scoring --output-dir [--num-examples ] [--visualize] +``` + +Required arguments: +- `--output-dir`: Path to the directory where scoring results will be saved + +Optional arguments: +- `--num-examples`: Number of examples to evaluate (useful for debugging) +- `--visualize`: If set, visualizes scoring results after completion +- `--scoring-params-config`: Path to the scoring params config file in JSON format +- `--dataset-id`: Pre-registered dataset_id to score (from llama-stack-client datasets list) +- `--dataset-path`: Path to the dataset file to score + + +## Eval Tasks +Manage evaluation tasks. + +### `llama-stack-client eval_tasks list` +Show available eval tasks on distribution endpoint +```bash +llama-stack-client eval_tasks list +``` + + +### `llama-stack-client eval_tasks register` +Register a new eval task +```bash +llama-stack-client eval_tasks register --eval-task-id --dataset-id --scoring-functions [--provider-id ] [--provider-eval-task-id ] [--metadata ] +``` + + +Required arguments: +- `--eval-task-id`: ID of the eval task +- `--dataset-id`: ID of the dataset to evaluate +- `--scoring-functions`: Scoring functions to use for evaluation + +Optional arguments: +- `--provider-id`: Provider ID for the eval task +- `--provider-eval-task-id`: Provider's eval task ID + + +## Tool Group Management +Manage available tool groups. + + +### `llama-stack-client toolgroups list` +Show available llama toolgroups at distribution endpoint +```bash +llama-stack-client toolgroups list +``` +``` ++---------------------------+------------------+------+---------------+ +| identifier | provider_id | args | mcp_endpoint | ++===========================+==================+======+===============+ +| builtin::rag | rag-runtime | None | None | ++---------------------------+------------------+------+---------------+ +| builtin::websearch | tavily-search | None | None | ++---------------------------+------------------+------+---------------+ +``` + +### `llama-stack-client toolgroups get` +Get available llama toolgroups by id +```bash +llama-stack-client toolgroups get +``` + +Shows detailed information about a specific toolgroup. If the toolgroup is not found, displays an error message. + + +Required arguments: +- `TOOLGROUP_ID`: ID of the tool group + + +### `llama-stack-client toolgroups register` +Register a new toolgroup at distribution endpoint +```bash +llama-stack-client toolgroups register [--provider-id ] [--provider-toolgroup-id ] [--mcp-config ] [--args ] +``` + + +Required arguments: +- `TOOLGROUP_ID`: ID of the tool group + +Optional arguments: +- `--provider-id`: Provider ID for the toolgroup +- `--provider-toolgroup-id`: Provider's toolgroup ID +- `--mcp-config`: JSON configuration for the MCP endpoint +- `--args`: JSON arguments for the toolgroup + +### `llama-stack-client toolgroups unregister` +Unregister a toolgroup from distribution endpoint +```bash +llama-stack-client toolgroups unregister +``` + + +Required arguments: +- `TOOLGROUP_ID`: ID of the tool group + + +## Datasets Management +Manage datasets. + + +### `llama-stack-client datasets list` +Show available datasets on distribution endpoint +```bash +llama-stack-client datasets list +``` + + +### `llama-stack-client datasets register` +```bash +llama-stack-client datasets register --dataset_id --purpose [--url ] [--dataset-id ] [--metadata ] +``` + +Required arguments: +- `--dataset_id`: Id of the dataset +- `--purpose`: Purpose of the dataset + +Optional arguments: +- `--metadata`: Metadata of the dataset +- `--url`: URL of the dataset +- `--dataset-path`: Local file path to the dataset. If specified, upload dataset via URL + + +### `llama-stack-client datasets unregister` +Remove a dataset +```bash +llama-stack-client datasets unregister +``` + + +Required arguments: +- `DATASET_ID`: Id of the dataset + + +## Scoring Functions Management +Manage scoring functions. + +### `llama-stack-client scoring_functions list` +Show available scoring functions on distribution endpoint +```bash +llama-stack-client scoring_functions list +``` +``` +โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ณโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”“ +โ”ƒ identifier โ”ƒ provider_id โ”ƒ description โ”ƒ type โ”ƒ +โ”กโ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ•‡โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”ฉ +โ”‚ basic::docvqa โ”‚ basic โ”‚ DocVQA Visual Question & Answer scoring function โ”‚ scoring_function โ”‚ +โ”‚ basic::equality โ”‚ basic โ”‚ Returns 1.0 if the input is equal to the target, 0.0 โ”‚ scoring_function โ”‚ +โ”‚ โ”‚ โ”‚ otherwise. โ”‚ โ”‚ +โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ +``` + + +### `llama-stack-client scoring_functions register` +Register a new scoring function +```bash +llama-stack-client scoring_functions register --scoring-fn-id --description --return-type [--provider-id ] [--provider-scoring-fn-id ] [--params ] +``` + + +Required arguments: +- `--scoring-fn-id`: Id of the scoring function +- `--description`: Description of the scoring function +- `--return-type`: Return type of the scoring function + +Optional arguments: +- `--provider-id`: Provider ID for the scoring function +- `--provider-scoring-fn-id`: Provider's scoring function ID +- `--params`: Parameters for the scoring function in JSON format + + +## Post Training Management +Post-training. + +### `llama-stack-client post_training list` +Show the list of available post training jobs +```bash +llama-stack-client post_training list +``` +```bash +["job-1", "job-2", "job-3"] +``` + + +### `llama-stack-client post_training artifacts` +Get the training artifacts of a specific post training job +```bash +llama-stack-client post_training artifacts --job-uuid +``` +```bash +JobArtifactsResponse(checkpoints=[], job_uuid='job-1') +``` + + +Required arguments: +- `--job-uuid`: Job UUID + + +### `llama-stack-client post_training supervised_fine_tune` +Kick off a supervised fine tune job +```bash +llama-stack-client post_training supervised_fine_tune --job-uuid --model --algorithm-config --training-config [--checkpoint-dir ] +``` + + +Required arguments: +- `--job-uuid`: Job UUID +- `--model`: Model ID +- `--algorithm-config`: Algorithm Config +- `--training-config`: Training Config + +Optional arguments: +- `--checkpoint-dir`: Checkpoint Config + + +### `llama-stack-client post_training status` +Show the status of a specific post training job +```bash +llama-stack-client post_training status --job-uuid +``` +```bash +JobStatusResponse( + checkpoints=[], + job_uuid='job-1', + status='completed', + completed_at="", + resources_allocated="", + scheduled_at="", + started_at="" +) +``` + + +Required arguments: +- `--job-uuid`: Job UUID + + +### `llama-stack-client post_training cancel` +Cancel the training job +```bash +llama-stack-client post_training cancel --job-uuid +``` +```bash +# This functionality is not yet implemented for llama-stack-client +โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ +โ”‚ Failed to post_training cancel_training_job โ”‚ +โ”‚ โ”‚ +โ”‚ Error Type: InternalServerError โ”‚ +โ”‚ Details: Error code: 501 - {'detail': 'Not implemented: '} โ”‚ +โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ +``` + + +Required arguments: +- `--job-uuid`: Job UUID diff --git a/versioned_docs/version-v0.2.23/references/python_sdk_reference/index.md b/versioned_docs/version-v0.2.23/references/python_sdk_reference/index.md new file mode 100644 index 0000000..b1a9396 --- /dev/null +++ b/versioned_docs/version-v0.2.23/references/python_sdk_reference/index.md @@ -0,0 +1,462 @@ +# Python SDK Reference + +## Shared Types + +```python +from llama_stack_client.types import ( + AgentConfig, + BatchCompletion, + CompletionMessage, + ContentDelta, + Document, + InterleavedContent, + InterleavedContentItem, + Message, + ParamType, + QueryConfig, + QueryResult, + ReturnType, + SafetyViolation, + SamplingParams, + ScoringResult, + SystemMessage, + ToolCall, + ToolParamDefinition, + ToolResponseMessage, + URL, + UserMessage, +) +``` + +## Toolgroups + +Types: + +```python +from llama_stack_client.types import ( + ListToolGroupsResponse, + ToolGroup, + ToolgroupListResponse, +) +``` + +Methods: + +- client.toolgroups.list() -> ToolgroupListResponse +- client.toolgroups.get(toolgroup_id) -> ToolGroup +- client.toolgroups.register(\*\*params) -> None +- client.toolgroups.unregister(toolgroup_id) -> None + +## Tools + +Types: + +```python +from llama_stack_client.types import ListToolsResponse, Tool, ToolListResponse +``` + +Methods: + +- client.tools.list(\*\*params) -> ToolListResponse +- client.tools.get(tool_name) -> Tool + +## ToolRuntime + +Types: + +```python +from llama_stack_client.types import ToolDef, ToolInvocationResult +``` + +Methods: + +- client.tool_runtime.invoke_tool(\*\*params) -> ToolInvocationResult +- client.tool_runtime.list_tools(\*\*params) -> JSONLDecoder[ToolDef] + +### RagTool + +Methods: + +- client.tool_runtime.rag_tool.insert(\*\*params) -> None +- client.tool_runtime.rag_tool.query(\*\*params) -> QueryResult + +## Agents + +Types: + +```python +from llama_stack_client.types import ( + InferenceStep, + MemoryRetrievalStep, + ShieldCallStep, + ToolExecutionStep, + ToolResponse, + AgentCreateResponse, +) +``` + +Methods: + +- client.agents.create(\*\*params) -> AgentCreateResponse +- client.agents.delete(agent_id) -> None + +### Session + +Types: + +```python +from llama_stack_client.types.agents import Session, SessionCreateResponse +``` + +Methods: + +- client.agents.session.create(agent_id, \*\*params) -> SessionCreateResponse +- client.agents.session.retrieve(session_id, \*, agent_id, \*\*params) -> Session +- client.agents.session.delete(session_id, \*, agent_id) -> None + +### Steps + +Types: + +```python +from llama_stack_client.types.agents import StepRetrieveResponse +``` + +Methods: + +- client.agents.steps.retrieve(step_id, \*, agent_id, session_id, turn_id) -> StepRetrieveResponse + +### Turn + +Types: + +```python +from llama_stack_client.types.agents import Turn, TurnCreateResponse +``` + +Methods: + +- client.agents.turn.create(session_id, \*, agent_id, \*\*params) -> TurnCreateResponse +- client.agents.turn.retrieve(turn_id, \*, agent_id, session_id) -> Turn + +## BatchInference + +Types: + +```python +from llama_stack_client.types import BatchInferenceChatCompletionResponse +``` + +Methods: + +- client.batch_inference.chat_completion(\*\*params) -> BatchInferenceChatCompletionResponse +- client.batch_inference.completion(\*\*params) -> BatchCompletion + +## Datasets + +Types: + +```python +from llama_stack_client.types import ( + ListDatasetsResponse, + DatasetRetrieveResponse, + DatasetListResponse, +) +``` + +Methods: + +- client.datasets.retrieve(dataset_id) -> Optional[DatasetRetrieveResponse] +- client.datasets.list() -> DatasetListResponse +- client.datasets.register(\*\*params) -> None +- client.datasets.unregister(dataset_id) -> None + +## Eval + +Types: + +```python +from llama_stack_client.types import EvaluateResponse, Job +``` + +Methods: + +- client.eval.evaluate_rows(benchmark_id, \*\*params) -> EvaluateResponse +- client.eval.run_eval(benchmark_id, \*\*params) -> Job + +### Jobs + +Types: + +```python +from llama_stack_client.types.eval import JobStatusResponse +``` + +Methods: + +- client.eval.jobs.retrieve(job_id, \*, benchmark_id) -> EvaluateResponse +- client.eval.jobs.cancel(job_id, \*, benchmark_id) -> None +- client.eval.jobs.status(job_id, \*, benchmark_id) -> Optional[JobStatusResponse] + +## Inspect + +Types: + +```python +from llama_stack_client.types import HealthInfo, ProviderInfo, RouteInfo, VersionInfo +``` + +Methods: + +- client.inspect.health() -> HealthInfo +- client.inspect.version() -> VersionInfo + +## Inference + +Types: + +```python +from llama_stack_client.types import ( + CompletionResponse, + EmbeddingsResponse, + TokenLogProbs, + InferenceChatCompletionResponse, + InferenceCompletionResponse, +) +``` + +Methods: + +- client.inference.chat_completion(\*\*params) -> InferenceChatCompletionResponse +- client.inference.completion(\*\*params) -> InferenceCompletionResponse +- client.inference.embeddings(\*\*params) -> EmbeddingsResponse + +## VectorIo + +Types: + +```python +from llama_stack_client.types import QueryChunksResponse +``` + +Methods: + +- client.vector_io.insert(\*\*params) -> None +- client.vector_io.query(\*\*params) -> QueryChunksResponse + +## VectorDBs + +Types: + +```python +from llama_stack_client.types import ( + ListVectorDBsResponse, + VectorDBRetrieveResponse, + VectorDBListResponse, + VectorDBRegisterResponse, +) +``` + +Methods: + +- client.vector_dbs.retrieve(vector_db_id) -> Optional[VectorDBRetrieveResponse] +- client.vector_dbs.list() -> VectorDBListResponse +- client.vector_dbs.register(\*\*params) -> VectorDBRegisterResponse +- client.vector_dbs.unregister(vector_db_id) -> None + +## Models + +Types: + +```python +from llama_stack_client.types import ListModelsResponse, Model, ModelListResponse +``` + +Methods: + +- client.models.retrieve(model_id) -> Optional[Model] +- client.models.list() -> ModelListResponse +- client.models.register(\*\*params) -> Model +- client.models.unregister(model_id) -> None + +## PostTraining + +Types: + +```python +from llama_stack_client.types import ListPostTrainingJobsResponse, PostTrainingJob +``` + +Methods: + +- client.post_training.preference_optimize(\*\*params) -> PostTrainingJob +- client.post_training.supervised_fine_tune(\*\*params) -> PostTrainingJob + +### Job + +Types: + +```python +from llama_stack_client.types.post_training import ( + JobListResponse, + JobArtifactsResponse, + JobStatusResponse, +) +``` + +Methods: + +- client.post_training.job.list() -> JobListResponse +- client.post_training.job.artifacts(\*\*params) -> Optional[JobArtifactsResponse] +- client.post_training.job.cancel(\*\*params) -> None +- client.post_training.job.status(\*\*params) -> Optional[JobStatusResponse] + +## Providers + +Types: + +```python +from llama_stack_client.types import ListProvidersResponse, ProviderListResponse +``` + +Methods: + +- client.providers.list() -> ProviderListResponse + +## Routes + +Types: + +```python +from llama_stack_client.types import ListRoutesResponse, RouteListResponse +``` + +Methods: + +- client.routes.list() -> RouteListResponse + +## Safety + +Types: + +```python +from llama_stack_client.types import RunShieldResponse +``` + +Methods: + +- client.safety.run_shield(\*\*params) -> RunShieldResponse + +## Shields + +Types: + +```python +from llama_stack_client.types import ListShieldsResponse, Shield, ShieldListResponse +``` + +Methods: + +- client.shields.retrieve(identifier) -> Optional[Shield] +- client.shields.list() -> ShieldListResponse +- client.shields.register(\*\*params) -> Shield + +## SyntheticDataGeneration + +Types: + +```python +from llama_stack_client.types import SyntheticDataGenerationResponse +``` + +Methods: + +- client.synthetic_data_generation.generate(\*\*params) -> SyntheticDataGenerationResponse + +## Telemetry + +Types: + +```python +from llama_stack_client.types import ( + QuerySpansResponse, + SpanWithStatus, + Trace, + TelemetryGetSpanResponse, + TelemetryGetSpanTreeResponse, + TelemetryQuerySpansResponse, + TelemetryQueryTracesResponse, +) +``` + +Methods: + +- client.telemetry.get_span(span_id, \*, trace_id) -> TelemetryGetSpanResponse +- client.telemetry.get_span_tree(span_id, \*\*params) -> TelemetryGetSpanTreeResponse +- client.telemetry.get_trace(trace_id) -> Trace +- client.telemetry.log_event(\*\*params) -> None +- client.telemetry.query_spans(\*\*params) -> TelemetryQuerySpansResponse +- client.telemetry.query_traces(\*\*params) -> TelemetryQueryTracesResponse +- client.telemetry.save_spans_to_dataset(\*\*params) -> None + +## Datasetio + +Types: + +```python +from llama_stack_client.types import PaginatedRowsResult +``` + +Methods: + +- client.datasetio.append_rows(\*\*params) -> None +- client.datasetio.get_rows_paginated(\*\*params) -> PaginatedRowsResult + +## Scoring + +Types: + +```python +from llama_stack_client.types import ScoringScoreResponse, ScoringScoreBatchResponse +``` + +Methods: + +- client.scoring.score(\*\*params) -> ScoringScoreResponse +- client.scoring.score_batch(\*\*params) -> ScoringScoreBatchResponse + +## ScoringFunctions + +Types: + +```python +from llama_stack_client.types import ( + ListScoringFunctionsResponse, + ScoringFn, + ScoringFunctionListResponse, +) +``` + +Methods: + +- client.scoring_functions.retrieve(scoring_fn_id) -> Optional[ScoringFn] +- client.scoring_functions.list() -> ScoringFunctionListResponse +- client.scoring_functions.register(\*\*params) -> None + +## Benchmarks + +Types: + +```python +from llama_stack_client.types import ( + Benchmark, + ListBenchmarksResponse, + BenchmarkListResponse, +) +``` + +Methods: + +- client.benchmarks.retrieve(benchmark_id) -> Optional[Benchmark] +- client.benchmarks.list() -> BenchmarkListResponse +- client.benchmarks.register(\*\*params) -> None diff --git a/versioned_sidebars/version-v0.2.23-sidebars.json b/versioned_sidebars/version-v0.2.23-sidebars.json new file mode 100644 index 0000000..5d5a219 --- /dev/null +++ b/versioned_sidebars/version-v0.2.23-sidebars.json @@ -0,0 +1,1371 @@ +{ + "tutorialSidebar": [ + "index", + { + "type": "category", + "label": "Getting Started", + "collapsed": false, + "items": [ + "getting_started/quickstart", + "getting_started/detailed_tutorial", + "getting_started/libraries" + ] + }, + { + "type": "category", + "label": "Concepts", + "collapsed": false, + "items": [ + "concepts/index", + "concepts/architecture", + { + "type": "category", + "label": "APIs", + "collapsed": true, + "items": [ + "concepts/apis/index", + "concepts/apis/api_providers", + "concepts/apis/external", + "concepts/apis/api_leveling" + ] + }, + "concepts/distributions", + "concepts/resources" + ] + }, + { + "type": "category", + "label": "Distributions", + "collapsed": false, + "items": [ + "distributions/index", + "distributions/list_of_distributions", + "distributions/building_distro", + "distributions/customizing_run_yaml", + "distributions/importing_as_library", + "distributions/configuration", + "distributions/starting_llama_stack_server", + { + "type": "category", + "label": "Self-Hosted Distributions", + "collapsed": true, + "items": [ + "distributions/self_hosted_distro/starter", + "distributions/self_hosted_distro/dell", + "distributions/self_hosted_distro/dell-tgi", + "distributions/self_hosted_distro/meta-reference-gpu", + "distributions/self_hosted_distro/nvidia", + "distributions/self_hosted_distro/passthrough" + ] + }, + { + "type": "category", + "label": "Remote-Hosted Distributions", + "collapsed": true, + "items": [ + "distributions/remote_hosted_distro/index", + "distributions/remote_hosted_distro/watsonx" + ] + }, + { + "type": "category", + "label": "On-Device Distributions", + "collapsed": true, + "items": [ + "distributions/ondevice_distro/ios_sdk", + "distributions/ondevice_distro/android_sdk" + ] + } + ] + }, + { + "type": "category", + "label": "Providers", + "collapsed": false, + "items": [ + "providers/index", + { + "type": "category", + "label": "Inference", + "collapsed": true, + "items": [ + "providers/inference/index", + "providers/inference/inline_meta-reference", + "providers/inference/inline_sentence-transformers", + "providers/inference/remote_anthropic", + "providers/inference/remote_azure", + "providers/inference/remote_bedrock", + "providers/inference/remote_cerebras", + "providers/inference/remote_databricks", + "providers/inference/remote_fireworks", + "providers/inference/remote_gemini", + "providers/inference/remote_groq", + "providers/inference/remote_hf_endpoint", + "providers/inference/remote_hf_serverless", + "providers/inference/remote_llama-openai-compat", + "providers/inference/remote_nvidia", + "providers/inference/remote_ollama", + "providers/inference/remote_openai", + "providers/inference/remote_passthrough", + "providers/inference/remote_runpod", + "providers/inference/remote_sambanova", + "providers/inference/remote_sambanova-openai-compat", + "providers/inference/remote_tgi", + "providers/inference/remote_together", + "providers/inference/remote_vertexai", + "providers/inference/remote_vllm", + "providers/inference/remote_watsonx" + ] + }, + { + "type": "category", + "label": "Safety", + "collapsed": true, + "items": [ + "providers/safety/index", + "providers/safety/inline_code-scanner", + "providers/safety/inline_llama-guard", + "providers/safety/inline_prompt-guard", + "providers/safety/remote_bedrock", + "providers/safety/remote_nvidia", + "providers/safety/remote_sambanova" + ] + }, + { + "type": "category", + "label": "Vector IO", + "collapsed": true, + "items": [ + "providers/vector_io/index", + "providers/vector_io/inline_chromadb", + "providers/vector_io/inline_faiss", + "providers/vector_io/inline_meta-reference", + "providers/vector_io/inline_milvus", + "providers/vector_io/inline_qdrant", + "providers/vector_io/inline_sqlite-vec", + "providers/vector_io/remote_chromadb", + "providers/vector_io/remote_milvus", + "providers/vector_io/remote_pgvector", + "providers/vector_io/remote_qdrant", + "providers/vector_io/remote_weaviate" + ] + }, + { + "type": "category", + "label": "Tool Runtime", + "collapsed": true, + "items": [ + "providers/tool_runtime/index", + "providers/tool_runtime/inline_rag-runtime", + "providers/tool_runtime/remote_bing-search", + "providers/tool_runtime/remote_brave-search", + "providers/tool_runtime/remote_model-context-protocol", + "providers/tool_runtime/remote_tavily-search", + "providers/tool_runtime/remote_wolfram-alpha" + ] + }, + { + "type": "category", + "label": "Agents", + "collapsed": true, + "items": [ + "providers/agents/index", + "providers/agents/inline_meta-reference" + ] + }, + { + "type": "category", + "label": "Post Training", + "collapsed": true, + "items": [ + "providers/post_training/index", + "providers/post_training/inline_huggingface", + "providers/post_training/inline_huggingface-cpu", + "providers/post_training/inline_huggingface-gpu", + "providers/post_training/inline_torchtune", + "providers/post_training/inline_torchtune-cpu", + "providers/post_training/inline_torchtune-gpu", + "providers/post_training/remote_nvidia" + ] + }, + { + "type": "category", + "label": "DatasetIO", + "collapsed": true, + "items": [ + "providers/datasetio/index", + "providers/datasetio/inline_localfs", + "providers/datasetio/remote_huggingface", + "providers/datasetio/remote_nvidia" + ] + }, + { + "type": "category", + "label": "Scoring", + "collapsed": true, + "items": [ + "providers/scoring/index", + "providers/scoring/inline_basic", + "providers/scoring/inline_braintrust", + "providers/scoring/inline_llm-as-judge" + ] + }, + { + "type": "category", + "label": "Files", + "collapsed": true, + "items": [ + "providers/files/index", + "providers/files/inline_localfs", + "providers/files/remote_s3" + ] + }, + { + "type": "category", + "label": "Eval", + "collapsed": true, + "items": [ + "providers/eval/index", + "providers/eval/inline_meta-reference", + "providers/eval/remote_nvidia" + ] + }, + { + "type": "category", + "label": "Telemetry", + "collapsed": true, + "items": [ + "providers/telemetry/index", + "providers/telemetry/inline_meta-reference" + ] + }, + { + "type": "category", + "label": "Batches", + "collapsed": true, + "items": [ + "providers/batches/index", + "providers/batches/inline_reference" + ] + }, + { + "type": "category", + "label": "External Providers", + "collapsed": true, + "items": [ + "providers/external/index", + "providers/external/external-providers-guide", + "providers/external/external-providers-list" + ] + }, + "providers/openai" + ] + }, + { + "type": "category", + "label": "Building Applications", + "collapsed": false, + "items": [ + "building_applications/index", + "building_applications/rag", + "building_applications/agent", + "building_applications/agent_execution_loop", + "building_applications/responses_vs_agents", + "building_applications/tools", + "building_applications/evals", + "building_applications/telemetry", + "building_applications/safety", + "building_applications/playground" + ] + }, + { + "type": "category", + "label": "Advanced APIs", + "collapsed": false, + "items": [ + "advanced_apis/post_training", + "advanced_apis/evaluation", + "advanced_apis/scoring" + ] + }, + { + "type": "category", + "label": "Deploying", + "collapsed": false, + "items": [ + "deploying/index", + "deploying/kubernetes_deployment", + "deploying/aws_eks_deployment" + ] + }, + { + "type": "category", + "label": "Contributing", + "collapsed": false, + "items": [ + "contributing/index", + "contributing/new_api_provider", + "contributing/new_vector_database", + "contributing/testing/record-replay" + ] + }, + { + "type": "category", + "label": "References", + "collapsed": false, + "items": [ + "references/index", + "references/llama_cli_reference/index", + "references/llama_stack_client_cli_reference", + "references/python_sdk_reference/index", + "references/evals_reference/index" + ] + } + ], + "apiSidebar": [ + { + "type": "doc", + "id": "api/llama-stack-specification" + }, + { + "type": "category", + "label": "Agents API for creating and interacting with agentic systems.", + "link": { + "type": "doc", + "id": "api/agents" + }, + "items": [ + { + "type": "doc", + "id": "api/list-all-agents", + "label": "List all agents.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/create-an-agent-with-the-given-configuration", + "label": "Create an agent with the given configuration.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/create-a-new-session-for-an-agent", + "label": "Create a new session for an agent.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/create-a-new-turn-for-an-agent", + "label": "Create a new turn for an agent.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/list-all-open-ai-responses", + "label": "List all OpenAI responses.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/create-a-new-open-ai-response", + "label": "Create a new OpenAI response.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/describe-an-agent-by-its-id", + "label": "Describe an agent by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/delete-an-agent-by-its-id-and-its-associated-sessions-and-turns", + "label": "Delete an agent by its ID and its associated sessions and turns.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/retrieve-an-agent-session-by-its-id", + "label": "Retrieve an agent session by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/delete-an-agent-session-by-its-id-and-its-associated-turns", + "label": "Delete an agent session by its ID and its associated turns.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/retrieve-an-open-ai-response-by-its-id", + "label": "Retrieve an OpenAI response by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/delete-an-open-ai-response-by-its-id", + "label": "Delete an OpenAI response by its ID.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/retrieve-an-agent-step-by-its-id", + "label": "Retrieve an agent step by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/retrieve-an-agent-turn-by-its-id", + "label": "Retrieve an agent turn by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/list-all-session-s-of-a-given-agent", + "label": "List all session(s) of a given agent.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/list-input-items-for-a-given-open-ai-response", + "label": "List input items for a given OpenAI response.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/resume-an-agent-turn-with-executed-tool-call-responses", + "label": "Resume an agent turn with executed tool call responses.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Batch inference API for generating completions and chat completions.", + "link": { + "type": "doc", + "id": "api/batch-inference-coming-soon" + }, + "items": [ + { + "type": "doc", + "id": "api/generate-a-chat-completion-for-the-given-messages-using-the-specified-model", + "label": "Generate a chat completion for the given messages using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/generate-a-completion-for-the-given-content-using-the-specified-model", + "label": "Generate a completion for the given content using the specified model.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Benchmarks", + "link": { + "type": "doc", + "id": "api/benchmarks" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-benchmark-by-its-id", + "label": "Get a benchmark by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-benchmark", + "label": "Unregister a benchmark.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/get-a-benchmark-by-its-id", + "label": "Get a benchmark by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-benchmark", + "label": "Unregister a benchmark.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-benchmarks", + "label": "List all benchmarks.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-benchmark", + "label": "Register a benchmark.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/list-all-benchmarks", + "label": "List all benchmarks.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-benchmark", + "label": "Register a benchmark.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "DatasetIO", + "link": { + "type": "doc", + "id": "api/dataset-io" + }, + "items": [ + { + "type": "doc", + "id": "api/append-rows-to-a-dataset", + "label": "Append rows to a dataset.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/get-a-paginated-list-of-rows-from-a-dataset", + "label": "Get a paginated list of rows from a dataset.", + "className": "api-method get" + } + ] + }, + { + "type": "category", + "label": "Datasets", + "link": { + "type": "doc", + "id": "api/datasets" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-dataset-by-its-id", + "label": "Get a dataset by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-dataset-by-its-id", + "label": "Unregister a dataset by its ID.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-datasets", + "label": "List all datasets.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-new-dataset", + "label": "Register a new dataset.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Llama Stack Evaluation API for running evaluations on model and agent candidates.", + "link": { + "type": "doc", + "id": "api/eval" + }, + "items": [ + { + "type": "doc", + "id": "api/evaluate-a-list-of-rows-on-a-benchmark", + "label": "Evaluate a list of rows on a benchmark.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/evaluate-a-list-of-rows-on-a-benchmark", + "label": "Evaluate a list of rows on a benchmark.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/get-the-status-of-a-job", + "label": "Get the status of a job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/cancel-a-job", + "label": "Cancel a job.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/get-the-status-of-a-job", + "label": "Get the status of a job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/cancel-a-job", + "label": "Cancel a job.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/get-the-result-of-a-job", + "label": "Get the result of a job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-the-result-of-a-job", + "label": "Get the result of a job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/run-an-evaluation-on-a-benchmark", + "label": "Run an evaluation on a benchmark.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/run-an-evaluation-on-a-benchmark", + "label": "Run an evaluation on a benchmark.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Files", + "link": { + "type": "doc", + "id": "api/files" + }, + "items": [ + { + "type": "doc", + "id": "api/returns-information-about-a-specific-file", + "label": "Returns information about a specific file.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/delete-a-file", + "label": "Delete a file.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/returns-a-list-of-files-that-belong-to-the-users-organization", + "label": "Returns a list of files that belong to the user's organization.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/upload-a-file-that-can-be-used-across-various-endpoints", + "label": "Upload a file that can be used across various endpoints.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/returns-the-contents-of-the-specified-file", + "label": "Returns the contents of the specified file.", + "className": "api-method get" + } + ] + }, + { + "type": "category", + "label": "Llama Stack Inference API for generating completions, chat completions, and embeddings.", + "link": { + "type": "doc", + "id": "api/inference" + }, + "items": [ + { + "type": "doc", + "id": "api/generate-chat-completions-for-a-batch-of-messages-using-the-specified-model", + "label": "Generate chat completions for a batch of messages using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/generate-completions-for-a-batch-of-content-using-the-specified-model", + "label": "Generate completions for a batch of content using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/generate-embeddings-for-content-pieces-using-the-specified-model", + "label": "Generate embeddings for content pieces using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/describe-a-chat-completion-by-its-id", + "label": "Describe a chat completion by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/list-all-chat-completions", + "label": "List all chat completions.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/generate-an-open-ai-compatible-chat-completion-for-the-given-messages-using-the-specified-model", + "label": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/generate-an-open-ai-compatible-completion-for-the-given-prompt-using-the-specified-model", + "label": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/generate-open-ai-compatible-embeddings-for-the-given-input-using-the-specified-model", + "label": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/rerank-a-list-of-documents-based-on-their-relevance-to-a-query", + "label": "Rerank a list of documents based on their relevance to a query.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Inspect", + "link": { + "type": "doc", + "id": "api/inspect" + }, + "items": [ + { + "type": "doc", + "id": "api/get-the-current-health-status-of-the-service", + "label": "Get the current health status of the service.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/list-all-available-api-routes-with-their-methods-and-implementing-providers", + "label": "List all available API routes with their methods and implementing providers.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-the-version-of-the-service", + "label": "Get the version of the service.", + "className": "api-method get" + } + ] + }, + { + "type": "category", + "label": "Models", + "link": { + "type": "doc", + "id": "api/models" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-model-by-its-identifier", + "label": "Get a model by its identifier.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-model", + "label": "Unregister a model.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-models", + "label": "List all models.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-model", + "label": "Register a model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/list-models-using-the-open-ai-api", + "label": "List models using the OpenAI API.", + "className": "api-method get" + } + ] + }, + { + "type": "category", + "label": "PostTraining (Coming Soon)", + "link": { + "type": "doc", + "id": "api/post-training-coming-soon" + }, + "items": [ + { + "type": "doc", + "id": "api/cancel-a-training-job", + "label": "Cancel a training job.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/cancel-a-training-job", + "label": "Cancel a training job.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/get-the-artifacts-of-a-training-job", + "label": "Get the artifacts of a training job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-the-artifacts-of-a-training-job", + "label": "Get the artifacts of a training job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-the-status-of-a-training-job", + "label": "Get the status of a training job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-the-status-of-a-training-job", + "label": "Get the status of a training job.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-all-training-jobs", + "label": "Get all training jobs.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-all-training-jobs", + "label": "Get all training jobs.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/run-preference-optimization-of-a-model", + "label": "Run preference optimization of a model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/run-preference-optimization-of-a-model", + "label": "Run preference optimization of a model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/run-supervised-fine-tuning-of-a-model", + "label": "Run supervised fine-tuning of a model.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/run-supervised-fine-tuning-of-a-model", + "label": "Run supervised fine-tuning of a model.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Protocol for prompt management operations.", + "link": { + "type": "doc", + "id": "api/prompts" + }, + "items": [ + { + "type": "doc", + "id": "api/list-all-prompts", + "label": "List all prompts.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/create-a-new-prompt", + "label": "Create a new prompt.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/get-a-prompt-by-its-identifier-and-optional-version", + "label": "Get a prompt by its identifier and optional version.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/update-an-existing-prompt-increments-version", + "label": "Update an existing prompt (increments version).", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/delete-a-prompt", + "label": "Delete a prompt.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-versions-of-a-specific-prompt", + "label": "List all versions of a specific prompt.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/set-which-version-of-a-prompt-should-be-the-default-in-get-prompt-latest", + "label": "Set which version of a prompt should be the default in get_prompt (latest).", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Providers API for inspecting, listing, and modifying providers and their configurations.", + "link": { + "type": "doc", + "id": "api/providers" + }, + "items": [ + { + "type": "doc", + "id": "api/get-detailed-information-about-a-specific-provider", + "label": "Get detailed information about a specific provider.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/list-all-available-providers", + "label": "List all available providers.", + "className": "api-method get" + } + ] + }, + { + "type": "category", + "label": "Safety", + "link": { + "type": "doc", + "id": "api/safety" + }, + "items": [ + { + "type": "doc", + "id": "api/classifies-if-text-and-or-image-inputs-are-potentially-harmful", + "label": "Classifies if text and/or image inputs are potentially harmful.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/run-a-shield", + "label": "Run a shield.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Scoring", + "link": { + "type": "doc", + "id": "api/scoring" + }, + "items": [ + { + "type": "doc", + "id": "api/score-a-list-of-rows", + "label": "Score a list of rows.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/score-a-batch-of-rows", + "label": "Score a batch of rows.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "ScoringFunctions", + "link": { + "type": "doc", + "id": "api/scoring-functions" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-scoring-function-by-its-id", + "label": "Get a scoring function by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-scoring-function", + "label": "Unregister a scoring function.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-scoring-functions", + "label": "List all scoring functions.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-scoring-function", + "label": "Register a scoring function.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Shields", + "link": { + "type": "doc", + "id": "api/shields" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-shield-by-its-identifier", + "label": "Get a shield by its identifier.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-shield", + "label": "Unregister a shield.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-shields", + "label": "List all shields.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-shield", + "label": "Register a shield.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "SyntheticDataGeneration (Coming Soon)", + "link": { + "type": "doc", + "id": "api/synthetic-data-generation-coming-soon" + }, + "items": [ + { + "type": "doc", + "id": "api/generate-synthetic-data-based-on-input-dialogs-and-apply-filtering", + "label": "Generate synthetic data based on input dialogs and apply filtering.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "Telemetry", + "link": { + "type": "doc", + "id": "api/telemetry" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-span-by-its-id", + "label": "Get a span by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-a-span-tree-by-its-id", + "label": "Get a span tree by its ID.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/get-a-trace-by-its-id", + "label": "Get a trace by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/log-an-event", + "label": "Log an event.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/query-metrics", + "label": "Query metrics.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/query-spans", + "label": "Query spans.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/query-traces", + "label": "Query traces.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/save-spans-to-a-dataset", + "label": "Save spans to a dataset.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "ToolGroups", + "link": { + "type": "doc", + "id": "api/tool-groups" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-tool-by-its-name", + "label": "Get a tool by its name.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/get-a-tool-group-by-its-id", + "label": "Get a tool group by its ID.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-tool-group", + "label": "Unregister a tool group.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-tool-groups-with-optional-provider", + "label": "List tool groups with optional provider.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-tool-group", + "label": "Register a tool group.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/list-tools-with-optional-tool-group", + "label": "List tools with optional tool group.", + "className": "api-method get" + } + ] + }, + { + "type": "category", + "label": "ToolRuntime", + "link": { + "type": "doc", + "id": "api/tool-runtime" + }, + "items": [ + { + "type": "doc", + "id": "api/index-documents-so-they-can-be-used-by-the-rag-system", + "label": "Index documents so they can be used by the RAG system.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/run-a-tool-with-the-given-arguments", + "label": "Run a tool with the given arguments.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/list-all-tools-in-the-runtime", + "label": "List all tools in the runtime.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/query-the-rag-system-for-context-typically-invoked-by-the-agent", + "label": "Query the RAG system for context; typically invoked by the agent.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "VectorDBs", + "link": { + "type": "doc", + "id": "api/vector-d-bs" + }, + "items": [ + { + "type": "doc", + "id": "api/get-a-vector-database-by-its-identifier", + "label": "Get a vector database by its identifier.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/unregister-a-vector-database", + "label": "Unregister a vector database.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/list-all-vector-databases", + "label": "List all vector databases.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/register-a-vector-database", + "label": "Register a vector database.", + "className": "api-method post" + } + ] + }, + { + "type": "category", + "label": "VectorIO", + "link": { + "type": "doc", + "id": "api/vector-io" + }, + "items": [ + { + "type": "doc", + "id": "api/insert-chunks-into-a-vector-database", + "label": "Insert chunks into a vector database.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/list-files-in-a-vector-store", + "label": "List files in a vector store.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/attach-a-file-to-a-vector-store", + "label": "Attach a file to a vector store.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/returns-a-list-of-vector-stores", + "label": "Returns a list of vector stores.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/creates-a-vector-store", + "label": "Creates a vector store.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/retrieves-a-vector-store", + "label": "Retrieves a vector store.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/updates-a-vector-store", + "label": "Updates a vector store.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/delete-a-vector-store", + "label": "Delete a vector store.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/retrieves-a-vector-store-file", + "label": "Retrieves a vector store file.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/updates-a-vector-store-file", + "label": "Updates a vector store file.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/delete-a-vector-store-file", + "label": "Delete a vector store file.", + "className": "api-method delete" + }, + { + "type": "doc", + "id": "api/retrieves-the-contents-of-a-vector-store-file", + "label": "Retrieves the contents of a vector store file.", + "className": "api-method get" + }, + { + "type": "doc", + "id": "api/search-for-chunks-in-a-vector-store", + "label": "Search for chunks in a vector store.", + "className": "api-method post" + }, + { + "type": "doc", + "id": "api/query-chunks-from-a-vector-database", + "label": "Query chunks from a vector database.", + "className": "api-method post" + } + ] + } + ] +} diff --git a/versions.json b/versions.json new file mode 100644 index 0000000..2a87029 --- /dev/null +++ b/versions.json @@ -0,0 +1,3 @@ +[ + "v0.2.23" +] diff --git a/versionsArchived.json b/versionsArchived.json new file mode 100644 index 0000000..295a056 --- /dev/null +++ b/versionsArchived.json @@ -0,0 +1,14 @@ +{ + "v0.2.22": "https://llamastack.github.io/v0.2.22/", + "v0.2.21": "https://llamastack.github.io/v0.2.21/", + "v0.2.20": "https://llamastack.github.io/v0.2.20/", + "v0.2.19": "https://llamastack.github.io/v0.2.19/", + "v0.2.18": "https://llamastack.github.io/v0.2.18/", + "v0.2.17": "https://llamastack.github.io/v0.2.17/", + "v0.2.16": "https://llamastack.github.io/v0.2.16/", + "v0.2.15": "https://llamastack.github.io/v0.2.15/", + "v0.2.14": "https://llamastack.github.io/v0.2.14/", + "v0.2.13": "https://llamastack.github.io/v0.2.13/", + "v0.2.12": "https://llamastack.github.io/v0.2.12/", + "v0.2.11": "https://llamastack.github.io/v0.2.11/" +}