From 38781e47b790990217a7a66792c2aaed82ae65cd Mon Sep 17 00:00:00 2001 From: Trevor Bergeron Date: Mon, 27 Apr 2026 18:56:14 +0000 Subject: [PATCH 1/2] test(bigframes): Use gemini-flash-2.5 in place of deprecated 2.0 --- .../apps/synthetic_data_generation.ipynb | 4 ++-- .../bq_dataframes_llm_code_generation.ipynb | 6 +++--- .../bq_dataframes_llm_kmeans.ipynb | 6 +++--- .../bq_dataframes_llm_vector_search.ipynb | 6 +++--- ...bq_dataframes_ml_drug_name_generation.ipynb | 4 ++-- .../generative_ai/large_language_models.ipynb | 4 ++-- .../bq_dataframes_template.ipynb | 4 ++-- ...-with-bigframes-over-national-jukebox.ipynb | 2 +- .../multimodal/multimodal_dataframe.ipynb | 4 ++-- .../tests/system/large/blob/test_function.py | 8 ++++---- .../tests/system/large/ml/test_llm.py | 18 ++++-------------- .../system/large/ml/test_multimodal_llm.py | 6 +++--- .../tests/system/large/operations/conftest.py | 2 +- .../bigframes/tests/system/load/test_llm.py | 6 +++--- 14 files changed, 35 insertions(+), 45 deletions(-) diff --git a/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb b/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb index b59777a5da3c..a5c3b281f0c4 100644 --- a/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb +++ b/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb @@ -98,7 +98,7 @@ }, { "cell_type": "code", - "execution_count": 3, + "execution_count": null, "metadata": { "colab": { "base_uri": "https://localhost:8080/", @@ -111,7 +111,7 @@ "source": [ "from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "model = GeminiTextGenerator(model_name=\"gemini-2.0-flash-001\")" + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb index c2c85be2b0cf..aa83d3ccc56d 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb @@ -29,7 +29,7 @@ "id": "JAPoU8Sm5E6e" }, "source": [ - "# Use BigQuery DataFrames with Generative AI for code generation", + "# Use BigQuery DataFrames with Generative AI for code generation\n", "\n", "\n", "\n", @@ -409,7 +409,7 @@ }, { "cell_type": "code", - "execution_count": 10, + "execution_count": null, "metadata": { "id": "sdjeXFwcHfl7" }, @@ -430,7 +430,7 @@ "source": [ "from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "model = GeminiTextGenerator(model_name=\"gemini-2.0-flash-001\")" + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb index 42dd5a99ac04..41e3d3e922ce 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb @@ -26,7 +26,7 @@ "cell_type": "markdown", "metadata": {}, "source": [ - "# Use BigQuery DataFrames to cluster and characterize complaints", + "# Use BigQuery DataFrames to cluster and characterize complaints\n", "\n", "
\n", "\n", @@ -1593,7 +1593,7 @@ }, { "cell_type": "code", - "execution_count": 19, + "execution_count": null, "metadata": { "id": "mL5P0_3X04dE" }, @@ -1614,7 +1614,7 @@ "source": [ "from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "q_a_model = GeminiTextGenerator(model_name=\"gemini-2.0-flash-001\")" + "q_a_model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb index 548162b2c6b3..63fee059646a 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb @@ -29,7 +29,7 @@ "id": "EQbZKS7_ooET" }, "source": [ - "# Build a Vector Search application using BigQuery DataFrames (aka BigFrames)", + "# Build a Vector Search application using BigQuery DataFrames (aka BigFrames)\n", "\n", "
\n", "\n", @@ -1451,7 +1451,7 @@ }, { "cell_type": "code", - "execution_count": 20, + "execution_count": null, "metadata": { "colab": { "base_uri": "https://localhost:8080/", @@ -1487,7 +1487,7 @@ "source": [ "## gemini model\n", "\n", - "llm_model = bf_llm.GeminiTextGenerator(model_name = \"gemini-2.0-flash-001\") ## replace with other model as needed" + "llm_model = bf_llm.GeminiTextGenerator(model_name = \"gemini-2.5-flash-001\") ## replace with other model as needed" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb index e51338c2e8fe..e5ac2651bd7b 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb @@ -480,7 +480,7 @@ }, { "cell_type": "code", - "execution_count": 22, + "execution_count": null, "metadata": { "id": "UW2fQ2k5Hsic" }, @@ -570,7 +570,7 @@ ], "source": [ "# Define the model\n", - "model = GeminiTextGenerator(model_name=\"gemini-2.0-flash-001\")\n", + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")\n", "\n", "# Invoke LLM with prompt\n", "response = predict(zero_shot_prompt, temperature = TEMPERATURE)\n", diff --git a/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb b/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb index e064394c4bd0..41cc5dfc0f79 100644 --- a/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb +++ b/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb @@ -21,7 +21,7 @@ }, { "cell_type": "code", - "execution_count": 3, + "execution_count": null, "metadata": {}, "outputs": [ { @@ -60,7 +60,7 @@ } ], "source": [ - "model = GeminiTextGenerator(model_name=\"gemini-2.0-flash-001\")" + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" ] }, { diff --git a/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb b/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb index 0970dcedc912..e125c89b2fd6 100644 --- a/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb +++ b/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb @@ -1305,13 +1305,13 @@ }, { "cell_type": "code", - "execution_count": 25, + "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "# model = GeminiTextGenerator(model_name=\"gemini-2.0-flash-001\")\n", + "# model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")\n", "\n", "# pred = model.predict(df)\n", "# pred" diff --git a/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb b/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb index fe2d567d1b31..f7e057223be6 100644 --- a/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb +++ b/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb @@ -521,7 +521,7 @@ "outputs": [], "source": [ "flattened[\"Transcription\"] = flattened[\"GCS Blob\"].blob.audio_transcribe(\n", - " model_name=\"gemini-2.0-flash-001\",\n", + " model_name=\"gemini-2.5-flash-001\",\n", " verbose=True,\n", ")\n", "flattened[\"Transcription\"]" diff --git a/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb b/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb index 1d3945c92c34..970d61d65ca5 100644 --- a/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb +++ b/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb @@ -1292,7 +1292,7 @@ }, { "cell_type": "code", - "execution_count": 18, + "execution_count": null, "metadata": {}, "outputs": [ { @@ -1339,7 +1339,7 @@ "\n", "transcribed_results = bbq.ai.generate(\n", " prompt=(prompt_text, audio_runtime),\n", - " endpoint=\"gemini-2.0-flash-001\",\n", + " endpoint=\"gemini-2.5-flash-001\",\n", " model_params={\"generationConfig\": {\"temperature\": 0.0}},\n", ")\n", "\n", diff --git a/packages/bigframes/tests/system/large/blob/test_function.py b/packages/bigframes/tests/system/large/blob/test_function.py index bc09baf268d1..eeb13188ff6b 100644 --- a/packages/bigframes/tests/system/large/blob/test_function.py +++ b/packages/bigframes/tests/system/large/blob/test_function.py @@ -754,8 +754,8 @@ def test_blob_pdf_chunk_verbose(pdf_mm_df: bpd.DataFrame, bq_connection: str): @pytest.mark.parametrize( "model_name", [ - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", + "gemini-2.5-flash-001", + "gemini-2.5-flash-lite-001", ], ) def test_blob_transcribe( @@ -805,8 +805,8 @@ def test_blob_transcribe( @pytest.mark.parametrize( "model_name", [ - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", + "gemini-2.5-flash-001", + "gemini-2.5-flash-lite-001", ], ) def test_blob_transcribe_verbose( diff --git a/packages/bigframes/tests/system/large/ml/test_llm.py b/packages/bigframes/tests/system/large/ml/test_llm.py index 357ac2d753ec..25718e92a949 100644 --- a/packages/bigframes/tests/system/large/ml/test_llm.py +++ b/packages/bigframes/tests/system/large/ml/test_llm.py @@ -27,8 +27,6 @@ @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", "gemini-2.5-pro", "gemini-2.5-flash", "gemini-2.5-flash-lite", @@ -56,8 +54,6 @@ def test_create_load_gemini_text_generator_model( @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", "gemini-2.5-pro", "gemini-2.5-flash", "gemini-2.5-flash-lite", @@ -79,8 +75,6 @@ def test_gemini_text_generator_predict_default_params_success( @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", "gemini-2.5-pro", "gemini-2.5-flash", "gemini-2.5-flash-lite", @@ -104,8 +98,6 @@ def test_gemini_text_generator_predict_with_params_success( @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", "gemini-2.5-pro", "gemini-2.5-flash", "gemini-2.5-flash-lite", @@ -131,8 +123,6 @@ def test_gemini_text_generator_multi_cols_predict_success( @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", "gemini-2.5-pro", "gemini-2.5-flash", "gemini-2.5-flash-lite", @@ -176,8 +166,8 @@ def test_gemini_text_generator_predict_output_schema_success( @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", + "gemini-2.5-flash-001", + "gemini-2.5-flash-lite-001", ), ) def test_llm_gemini_score(llm_fine_tune_df_default_index, model_name): @@ -205,8 +195,8 @@ def test_llm_gemini_score(llm_fine_tune_df_default_index, model_name): @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", + "gemini-2.5-flash-001", + "gemini-2.5-flash-lite-001", ), ) def test_llm_gemini_pro_score_params(llm_fine_tune_df_default_index, model_name): diff --git a/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py b/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py index 69c316e3dacc..1aead537629d 100644 --- a/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py +++ b/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py @@ -24,8 +24,8 @@ @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", + "gemini-2.5-flash-001", + "gemini-2.5-flash-lite-001", ), ) @pytest.mark.flaky(retries=2) @@ -65,7 +65,7 @@ def test_multimodal_embedding_generator_predict_default_params_success( @pytest.mark.parametrize( "model_name", - ("gemini-2.0-flash-001",), + ("gemini-2.5-flash-001",), ) @pytest.mark.flaky(retries=2) def test_gemini_text_generator_multimodal_structured_output( diff --git a/packages/bigframes/tests/system/large/operations/conftest.py b/packages/bigframes/tests/system/large/operations/conftest.py index 6f64c7552f32..74e16bc0d56e 100644 --- a/packages/bigframes/tests/system/large/operations/conftest.py +++ b/packages/bigframes/tests/system/large/operations/conftest.py @@ -22,7 +22,7 @@ def gemini_flash_model(session, bq_connection) -> llm.GeminiTextGenerator: return llm.GeminiTextGenerator( session=session, connection_name=bq_connection, - model_name="gemini-2.0-flash-001", + model_name="gemini-2.5-flash-001", ) diff --git a/packages/bigframes/tests/system/load/test_llm.py b/packages/bigframes/tests/system/load/test_llm.py index 4e8bc65912fb..be848edcbc6c 100644 --- a/packages/bigframes/tests/system/load/test_llm.py +++ b/packages/bigframes/tests/system/load/test_llm.py @@ -41,8 +41,8 @@ def llm_remote_text_df(session, llm_remote_text_pandas_df): @pytest.mark.parametrize( "model_name", ( - "gemini-2.0-flash-001", - "gemini-2.0-flash-lite-001", + "gemini-2.5-flash-001", + "gemini-2.5-flash-lite-001", ), ) def test_llm_gemini_configure_fit( @@ -79,7 +79,7 @@ def test_llm_gemini_configure_fit( @pytest.mark.flaky(retries=2) def test_llm_gemini_w_ground_with_google_search(llm_remote_text_df): - model = llm.GeminiTextGenerator(model_name="gemini-2.0-flash-001", max_iterations=1) + model = llm.GeminiTextGenerator(model_name="gemini-2.5-flash-001", max_iterations=1) df = model.predict( llm_remote_text_df["prompt"], ground_with_google_search=True, From f55aef56fe470bad2a9f283de054818ac63b7961 Mon Sep 17 00:00:00 2001 From: Trevor Bergeron Date: Mon, 27 Apr 2026 19:40:15 +0000 Subject: [PATCH 2/2] remove 001 suffix --- .../notebooks/apps/synthetic_data_generation.ipynb | 2 +- .../generative_ai/bq_dataframes_llm_code_generation.ipynb | 2 +- .../generative_ai/bq_dataframes_llm_kmeans.ipynb | 2 +- .../generative_ai/bq_dataframes_llm_vector_search.ipynb | 2 +- .../bq_dataframes_ml_drug_name_generation.ipynb | 2 +- .../notebooks/generative_ai/large_language_models.ipynb | 2 +- .../getting_started/bq_dataframes_template.ipynb | 2 +- ...ctor-search-with-bigframes-over-national-jukebox.ipynb | 2 +- .../notebooks/multimodal/multimodal_dataframe.ipynb | 2 +- .../bigframes/tests/system/large/blob/test_function.py | 8 ++++---- packages/bigframes/tests/system/large/ml/test_llm.py | 8 ++++---- .../tests/system/large/ml/test_multimodal_llm.py | 6 +++--- .../bigframes/tests/system/large/operations/conftest.py | 2 +- packages/bigframes/tests/system/load/test_llm.py | 6 +++--- 14 files changed, 24 insertions(+), 24 deletions(-) diff --git a/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb b/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb index a5c3b281f0c4..00d30fc8a8a9 100644 --- a/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb +++ b/packages/bigframes/notebooks/apps/synthetic_data_generation.ipynb @@ -111,7 +111,7 @@ "source": [ "from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash\")" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb index aa83d3ccc56d..527d3c4aaacd 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_code_generation.ipynb @@ -430,7 +430,7 @@ "source": [ "from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash\")" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb index 41e3d3e922ce..2d5bb46d95ed 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_kmeans.ipynb @@ -1614,7 +1614,7 @@ "source": [ "from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "q_a_model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" + "q_a_model = GeminiTextGenerator(model_name=\"gemini-2.5-flash\")" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb index 63fee059646a..c9fa39926a92 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_llm_vector_search.ipynb @@ -1487,7 +1487,7 @@ "source": [ "## gemini model\n", "\n", - "llm_model = bf_llm.GeminiTextGenerator(model_name = \"gemini-2.5-flash-001\") ## replace with other model as needed" + "llm_model = bf_llm.GeminiTextGenerator(model_name = \"gemini-2.5-flash\") ## replace with other model as needed" ] }, { diff --git a/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb b/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb index e5ac2651bd7b..93ac3f31c140 100644 --- a/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb +++ b/packages/bigframes/notebooks/generative_ai/bq_dataframes_ml_drug_name_generation.ipynb @@ -570,7 +570,7 @@ ], "source": [ "# Define the model\n", - "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")\n", + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash\")\n", "\n", "# Invoke LLM with prompt\n", "response = predict(zero_shot_prompt, temperature = TEMPERATURE)\n", diff --git a/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb b/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb index 41cc5dfc0f79..4ff9a9d3d237 100644 --- a/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb +++ b/packages/bigframes/notebooks/generative_ai/large_language_models.ipynb @@ -60,7 +60,7 @@ } ], "source": [ - "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")" + "model = GeminiTextGenerator(model_name=\"gemini-2.5-flash\")" ] }, { diff --git a/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb b/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb index e125c89b2fd6..664a3a68d33a 100644 --- a/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb +++ b/packages/bigframes/notebooks/getting_started/bq_dataframes_template.ipynb @@ -1311,7 +1311,7 @@ "source": [ "# from bigframes.ml.llm import GeminiTextGenerator\n", "\n", - "# model = GeminiTextGenerator(model_name=\"gemini-2.5-flash-001\")\n", + "# model = GeminiTextGenerator(model_name=\"gemini-2.5-flash\")\n", "\n", "# pred = model.predict(df)\n", "# pred" diff --git a/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb b/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb index f7e057223be6..4faff4b8e768 100644 --- a/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb +++ b/packages/bigframes/notebooks/kaggle/vector-search-with-bigframes-over-national-jukebox.ipynb @@ -521,7 +521,7 @@ "outputs": [], "source": [ "flattened[\"Transcription\"] = flattened[\"GCS Blob\"].blob.audio_transcribe(\n", - " model_name=\"gemini-2.5-flash-001\",\n", + " model_name=\"gemini-2.5-flash\",\n", " verbose=True,\n", ")\n", "flattened[\"Transcription\"]" diff --git a/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb b/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb index 970d61d65ca5..8f3241259d5f 100644 --- a/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb +++ b/packages/bigframes/notebooks/multimodal/multimodal_dataframe.ipynb @@ -1339,7 +1339,7 @@ "\n", "transcribed_results = bbq.ai.generate(\n", " prompt=(prompt_text, audio_runtime),\n", - " endpoint=\"gemini-2.5-flash-001\",\n", + " endpoint=\"gemini-2.5-flash\",\n", " model_params={\"generationConfig\": {\"temperature\": 0.0}},\n", ")\n", "\n", diff --git a/packages/bigframes/tests/system/large/blob/test_function.py b/packages/bigframes/tests/system/large/blob/test_function.py index eeb13188ff6b..e0996db4212a 100644 --- a/packages/bigframes/tests/system/large/blob/test_function.py +++ b/packages/bigframes/tests/system/large/blob/test_function.py @@ -754,8 +754,8 @@ def test_blob_pdf_chunk_verbose(pdf_mm_df: bpd.DataFrame, bq_connection: str): @pytest.mark.parametrize( "model_name", [ - "gemini-2.5-flash-001", - "gemini-2.5-flash-lite-001", + "gemini-2.5-flash", + "gemini-2.5-flash-lite", ], ) def test_blob_transcribe( @@ -805,8 +805,8 @@ def test_blob_transcribe( @pytest.mark.parametrize( "model_name", [ - "gemini-2.5-flash-001", - "gemini-2.5-flash-lite-001", + "gemini-2.5-flash", + "gemini-2.5-flash-lite", ], ) def test_blob_transcribe_verbose( diff --git a/packages/bigframes/tests/system/large/ml/test_llm.py b/packages/bigframes/tests/system/large/ml/test_llm.py index 25718e92a949..638e151ca141 100644 --- a/packages/bigframes/tests/system/large/ml/test_llm.py +++ b/packages/bigframes/tests/system/large/ml/test_llm.py @@ -166,8 +166,8 @@ def test_gemini_text_generator_predict_output_schema_success( @pytest.mark.parametrize( "model_name", ( - "gemini-2.5-flash-001", - "gemini-2.5-flash-lite-001", + "gemini-2.5-flash", + "gemini-2.5-flash-lite", ), ) def test_llm_gemini_score(llm_fine_tune_df_default_index, model_name): @@ -195,8 +195,8 @@ def test_llm_gemini_score(llm_fine_tune_df_default_index, model_name): @pytest.mark.parametrize( "model_name", ( - "gemini-2.5-flash-001", - "gemini-2.5-flash-lite-001", + "gemini-2.5-flash", + "gemini-2.5-flash-lite", ), ) def test_llm_gemini_pro_score_params(llm_fine_tune_df_default_index, model_name): diff --git a/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py b/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py index 1aead537629d..6babac261f3c 100644 --- a/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py +++ b/packages/bigframes/tests/system/large/ml/test_multimodal_llm.py @@ -24,8 +24,8 @@ @pytest.mark.parametrize( "model_name", ( - "gemini-2.5-flash-001", - "gemini-2.5-flash-lite-001", + "gemini-2.5-flash", + "gemini-2.5-flash-lite", ), ) @pytest.mark.flaky(retries=2) @@ -65,7 +65,7 @@ def test_multimodal_embedding_generator_predict_default_params_success( @pytest.mark.parametrize( "model_name", - ("gemini-2.5-flash-001",), + ("gemini-2.5-flash",), ) @pytest.mark.flaky(retries=2) def test_gemini_text_generator_multimodal_structured_output( diff --git a/packages/bigframes/tests/system/large/operations/conftest.py b/packages/bigframes/tests/system/large/operations/conftest.py index 74e16bc0d56e..0122c860ca74 100644 --- a/packages/bigframes/tests/system/large/operations/conftest.py +++ b/packages/bigframes/tests/system/large/operations/conftest.py @@ -22,7 +22,7 @@ def gemini_flash_model(session, bq_connection) -> llm.GeminiTextGenerator: return llm.GeminiTextGenerator( session=session, connection_name=bq_connection, - model_name="gemini-2.5-flash-001", + model_name="gemini-2.5-flash", ) diff --git a/packages/bigframes/tests/system/load/test_llm.py b/packages/bigframes/tests/system/load/test_llm.py index be848edcbc6c..eec76cf9b679 100644 --- a/packages/bigframes/tests/system/load/test_llm.py +++ b/packages/bigframes/tests/system/load/test_llm.py @@ -41,8 +41,8 @@ def llm_remote_text_df(session, llm_remote_text_pandas_df): @pytest.mark.parametrize( "model_name", ( - "gemini-2.5-flash-001", - "gemini-2.5-flash-lite-001", + "gemini-2.5-flash", + "gemini-2.5-flash-lite", ), ) def test_llm_gemini_configure_fit( @@ -79,7 +79,7 @@ def test_llm_gemini_configure_fit( @pytest.mark.flaky(retries=2) def test_llm_gemini_w_ground_with_google_search(llm_remote_text_df): - model = llm.GeminiTextGenerator(model_name="gemini-2.5-flash-001", max_iterations=1) + model = llm.GeminiTextGenerator(model_name="gemini-2.5-flash", max_iterations=1) df = model.predict( llm_remote_text_df["prompt"], ground_with_google_search=True,