Skip to content

Commit

Permalink
feat(generative_ai): add model evaluation sample (#11122)
Browse files Browse the repository at this point in the history
* feat(generative_ai): add model evaluation sample

* set credentials required by pipeline components
  • Loading branch information
gericdong committed Jan 15, 2024
1 parent 93429e0 commit 583de9f
Show file tree
Hide file tree
Showing 2 changed files with 93 additions and 0 deletions.
59 changes: 59 additions & 0 deletions generative_ai/evaluate_model.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,59 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

# [START aiplatform_evaluate_model]

from google.auth import default
import vertexai
from vertexai.preview.language_models import (
EvaluationTextClassificationSpec,
TextGenerationModel,
)

# Set credentials for the pipeline components used in the evaluation task
credentials, _ = default(scopes=["https://www.googleapis.com/auth/cloud-platform"])


def evaluate_model(
project_id: str,
location: str,
) -> object:
"""Evaluate the performance of a generative AI model."""

vertexai.init(project=project_id, location=location, credentials=credentials)

# Create a reference to a generative AI model
model = TextGenerationModel.from_pretrained("text-bison@001")

# Define the evaluation specification for a text classification task
task_spec = EvaluationTextClassificationSpec(
ground_truth_data=[
"gs://cloud-samples-data/ai-platform/generative_ai/llm_classification_bp_input_prompts_with_ground_truth.jsonl"
],
class_names=["nature", "news", "sports", "health", "startups"],
target_column_name="ground_truth",
)

# Evaluate the model
eval_metrics = model.evaluate(task_spec=task_spec)
print(eval_metrics)

return eval_metrics


# [END aiplatform_evaluate_model]


if __name__ == "__main__":
evaluate_model()
34 changes: 34 additions & 0 deletions generative_ai/evaluate_model_test.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,34 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import os

import backoff
from google.api_core.exceptions import ResourceExhausted

import evaluate_model


_PROJECT_ID = os.getenv("GOOGLE_CLOUD_PROJECT")
_LOCATION = "us-central1"


@backoff.on_exception(backoff.expo, ResourceExhausted, max_time=10)
def test_evaluate_model() -> None:
eval_metrics = evaluate_model.evaluate_model(
_PROJECT_ID,
_LOCATION,
)

assert hasattr(eval_metrics, "auRoc")

0 comments on commit 583de9f

Please sign in to comment.