diff --git a/packages/sdk/server-ai/examples/vercel-ai/package.json b/packages/sdk/server-ai/examples/vercel-ai/package.json index c2d9ad107..ef07a631d 100644 --- a/packages/sdk/server-ai/examples/vercel-ai/package.json +++ b/packages/sdk/server-ai/examples/vercel-ai/package.json @@ -25,7 +25,7 @@ "@launchdarkly/node-server-sdk": "^9.7.1", "@launchdarkly/server-sdk-ai": "^0.15.0", "@launchdarkly/server-sdk-ai-vercel": "^0.4.0", - "ai": "5.0.0", + "ai": "5.0.52", "zod": "^3.23.8" }, "devDependencies": { diff --git a/packages/sdk/server-ai/examples/vercel-ai/src/index.ts b/packages/sdk/server-ai/examples/vercel-ai/src/index.ts index 9ae9b6144..badac54e4 100644 --- a/packages/sdk/server-ai/examples/vercel-ai/src/index.ts +++ b/packages/sdk/server-ai/examples/vercel-ai/src/index.ts @@ -68,7 +68,7 @@ async function main() { // Call the model and track metrics for the ai config const result = await aiConfig.tracker.trackMetricsOf( VercelProvider.getAIMetricsFromResponse, - () => generateText(vercelConfig), + () => generateText({ ...vercelConfig, messages: vercelConfig.messages ?? [] }), ); console.log('Response:', result.text); @@ -92,7 +92,7 @@ async function main() { // Stream is returned immediately (synchronously), metrics tracked in background const streamResult = aiConfig.tracker.trackStreamMetricsOf( - () => streamText(vercelConfig), + () => streamText({ ...vercelConfig, messages: vercelConfig.messages ?? [] }), VercelProvider.getAIMetricsFromStream, );