diff --git a/docs/dist/examples/generate_structured_data_cohere.ipynb b/docs/dist/examples/generate_structured_data_cohere.ipynb index b5f72cb20..22007ad92 100644 --- a/docs/dist/examples/generate_structured_data_cohere.ipynb +++ b/docs/dist/examples/generate_structured_data_cohere.ipynb @@ -146,7 +146,7 @@ }, { "cell_type": "code", - "execution_count": 6, + "execution_count": null, "id": "42766922-14d0-4b5e-853a-23f05b896a09", "metadata": {}, "outputs": [ @@ -171,7 +171,7 @@ "\n", "raw_llm_response, validated_response, *rest = guard(\n", " messages=[{\"role\": \"user\", \"content\": prompt}],\n", - " model=\"command-r\",\n", + " model=\"command-r-08-2024\",\n", " max_tokens=1024,\n", " temperature=0.3,\n", ")" diff --git a/docs/dist/examples/generate_structured_data_cohere.md b/docs/dist/examples/generate_structured_data_cohere.md index 5a6d2f7ef..361feda90 100644 --- a/docs/dist/examples/generate_structured_data_cohere.md +++ b/docs/dist/examples/generate_structured_data_cohere.md @@ -105,7 +105,7 @@ guard = gd.Guard.for_pydantic(output_class=Orders) raw_llm_response, validated_response, *rest = guard( messages=[{"role": "user", "content": prompt}], - model="command-r", + model="command-r-08-2024", max_tokens=1024, temperature=0.3, ) diff --git a/docs/src/examples/generate_structured_data_cohere.ipynb b/docs/src/examples/generate_structured_data_cohere.ipynb index b5f72cb20..22007ad92 100644 --- a/docs/src/examples/generate_structured_data_cohere.ipynb +++ b/docs/src/examples/generate_structured_data_cohere.ipynb @@ -146,7 +146,7 @@ }, { "cell_type": "code", - "execution_count": 6, + "execution_count": null, "id": "42766922-14d0-4b5e-853a-23f05b896a09", "metadata": {}, "outputs": [ @@ -171,7 +171,7 @@ "\n", "raw_llm_response, validated_response, *rest = guard(\n", " messages=[{\"role\": \"user\", \"content\": prompt}],\n", - " model=\"command-r\",\n", + " model=\"command-r-08-2024\",\n", " max_tokens=1024,\n", " temperature=0.3,\n", ")"