diff --git a/packages/uipath/src/uipath/eval/evaluators/legacy_context_precision_evaluator.py b/packages/uipath/src/uipath/eval/evaluators/legacy_context_precision_evaluator.py index ec47d2c62..3b0468363 100644 --- a/packages/uipath/src/uipath/eval/evaluators/legacy_context_precision_evaluator.py +++ b/packages/uipath/src/uipath/eval/evaluators/legacy_context_precision_evaluator.py @@ -4,6 +4,8 @@ import json from typing import Any, Optional +from pydantic import Field + from uipath.platform import UiPath from uipath.platform.chat import UiPathLlmChatService @@ -99,7 +101,7 @@ class LegacyContextPrecisionEvaluator( The final score is the mean of all chunk relevancy scores (normalized to 0-1). """ - model: str + model: str = Field(default="same-as-agent") query_placeholder: str = "{{Query}}" chunks_placeholder: str = "{{Chunks}}" llm: Optional[UiPathLlmChatService] = None diff --git a/packages/uipath/src/uipath/eval/evaluators/legacy_faithfulness_evaluator.py b/packages/uipath/src/uipath/eval/evaluators/legacy_faithfulness_evaluator.py index 5cd30c9db..c4eba1fcf 100644 --- a/packages/uipath/src/uipath/eval/evaluators/legacy_faithfulness_evaluator.py +++ b/packages/uipath/src/uipath/eval/evaluators/legacy_faithfulness_evaluator.py @@ -3,6 +3,8 @@ import json from typing import Any, Optional +from pydantic import Field + from uipath.platform import UiPath from uipath.platform.chat import UiPathLlmChatService @@ -39,7 +41,7 @@ class LegacyFaithfulnessEvaluator( The final score is the percentage of claims that are grounded. """ - model: str + model: str = Field(default="same-as-agent") llm: Optional[UiPathLlmChatService] = None def model_post_init(self, __context: Any): diff --git a/packages/uipath/src/uipath/eval/evaluators/legacy_llm_as_judge_evaluator.py b/packages/uipath/src/uipath/eval/evaluators/legacy_llm_as_judge_evaluator.py index 5b9a8ee1c..0c676ac9b 100644 --- a/packages/uipath/src/uipath/eval/evaluators/legacy_llm_as_judge_evaluator.py +++ b/packages/uipath/src/uipath/eval/evaluators/legacy_llm_as_judge_evaluator.py @@ -3,7 +3,7 @@ import logging from typing import Any, Optional -from pydantic import field_validator +from pydantic import Field, field_validator from uipath.platform import UiPath from uipath.platform.chat import UiPathLlmChatService @@ -41,7 +41,7 @@ class LegacyLlmAsAJudgeEvaluator(BaseLegacyEvaluator[LegacyLlmAsAJudgeEvaluatorC """Legacy evaluator that uses an LLM to judge the quality of agent output.""" prompt: str - model: str + model: str = Field(default="same-as-agent") actual_output_placeholder: str = "{{ActualOutput}}" expected_output_placeholder: str = "{{ExpectedOutput}}" llm: Optional[UiPathLlmChatService] = None diff --git a/packages/uipath/src/uipath/eval/evaluators/legacy_trajectory_evaluator.py b/packages/uipath/src/uipath/eval/evaluators/legacy_trajectory_evaluator.py index 24d56560b..17b69d0d0 100644 --- a/packages/uipath/src/uipath/eval/evaluators/legacy_trajectory_evaluator.py +++ b/packages/uipath/src/uipath/eval/evaluators/legacy_trajectory_evaluator.py @@ -4,7 +4,7 @@ from typing import Any, Optional from opentelemetry.sdk.trace import ReadableSpan -from pydantic import field_validator +from pydantic import Field, field_validator from uipath.platform import UiPath from uipath.platform.chat import UiPathLlmChatService @@ -42,7 +42,7 @@ class LegacyTrajectoryEvaluator(BaseLegacyEvaluator[LegacyTrajectoryEvaluatorCon """Legacy evaluator that analyzes the trajectory/path taken to reach outputs.""" prompt: str - model: str + model: str = Field(default="same-as-agent") expected_agent_behavior_placeholder: str = "{{ExpectedAgentBehavior}}" agent_run_history_placeholder: str = "{{AgentRunHistory}}" llm: Optional[UiPathLlmChatService] = None