From 095abe62eb9a1c7931da3e48438d3278f6991eb0 Mon Sep 17 00:00:00 2001 From: Wauplin Date: Tue, 17 Jun 2025 09:48:40 +0200 Subject: [PATCH] [InferenceSnippets] Fix HF_TOKEN not used if https:// in example --- packages/inference/src/snippets/getInferenceSnippets.ts | 2 +- .../js/huggingface.js/0.fireworks-ai.js | 2 +- .../python/huggingface_hub/0.fireworks-ai.py | 2 +- .../js/huggingface.js/0.fireworks-ai.js | 2 +- .../python/huggingface_hub/0.fireworks-ai.py | 2 +- 5 files changed, 5 insertions(+), 5 deletions(-) diff --git a/packages/inference/src/snippets/getInferenceSnippets.ts b/packages/inference/src/snippets/getInferenceSnippets.ts index 96377ed8b1..077bc73fb1 100644 --- a/packages/inference/src/snippets/getInferenceSnippets.ts +++ b/packages/inference/src/snippets/getInferenceSnippets.ts @@ -466,7 +466,7 @@ function replaceAccessTokenPlaceholder( !endpointUrl && // custom endpointUrl => use a generic API_TOKEN (provider == "hf-inference" || // hf-inference provider => use $HF_TOKEN (!directRequest && // if explicit directRequest => use provider-specific token - (!snippet.includes("https://") || // no URL provided => using a client => use $HF_TOKEN + (snippet.includes("InferenceClient") || // using a client => use $HF_TOKEN snippet.includes("https://router.huggingface.co")))); // explicit routed request => use $HF_TOKEN const accessTokenEnvVar = useHfToken ? "HF_TOKEN" // e.g. routed request or hf-inference diff --git a/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/js/huggingface.js/0.fireworks-ai.js b/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/js/huggingface.js/0.fireworks-ai.js index 08252975cc..e7de7159de 100644 --- a/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/js/huggingface.js/0.fireworks-ai.js +++ b/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/js/huggingface.js/0.fireworks-ai.js @@ -1,6 +1,6 @@ import { InferenceClient } from "@huggingface/inference"; -const client = new InferenceClient(process.env.FIREWORKS_AI_API_KEY); +const client = new InferenceClient(process.env.HF_TOKEN); const chatCompletion = await client.chatCompletion({ provider: "fireworks-ai", diff --git a/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/python/huggingface_hub/0.fireworks-ai.py b/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/python/huggingface_hub/0.fireworks-ai.py index 7c228773ae..7dbcbf5bb7 100644 --- a/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/python/huggingface_hub/0.fireworks-ai.py +++ b/packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/python/huggingface_hub/0.fireworks-ai.py @@ -3,7 +3,7 @@ client = InferenceClient( provider="fireworks-ai", - api_key=os.environ["FIREWORKS_AI_API_KEY"], + api_key=os.environ["HF_TOKEN"], ) completion = client.chat.completions.create( diff --git a/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/js/huggingface.js/0.fireworks-ai.js b/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/js/huggingface.js/0.fireworks-ai.js index 0565060cac..109d9ca82b 100644 --- a/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/js/huggingface.js/0.fireworks-ai.js +++ b/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/js/huggingface.js/0.fireworks-ai.js @@ -1,6 +1,6 @@ import { InferenceClient } from "@huggingface/inference"; -const client = new InferenceClient(process.env.FIREWORKS_AI_API_KEY); +const client = new InferenceClient(process.env.HF_TOKEN); let out = ""; diff --git a/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/python/huggingface_hub/0.fireworks-ai.py b/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/python/huggingface_hub/0.fireworks-ai.py index 4cae631fe2..bef166d6aa 100644 --- a/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/python/huggingface_hub/0.fireworks-ai.py +++ b/packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/python/huggingface_hub/0.fireworks-ai.py @@ -3,7 +3,7 @@ client = InferenceClient( provider="fireworks-ai", - api_key=os.environ["FIREWORKS_AI_API_KEY"], + api_key=os.environ["HF_TOKEN"], ) stream = client.chat.completions.create(