diff --git a/pyproject.toml b/pyproject.toml index 6ed0f997b..d1c16b85e 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "uipath" -version = "2.7.9" +version = "2.7.10" description = "Python SDK and CLI for UiPath Platform, enabling programmatic interaction with automation services, process management, and deployment tools." readme = { file = "README.md", content-type = "text/markdown" } requires-python = ">=3.11" diff --git a/src/uipath/eval/evaluators/llm_as_judge_evaluator.py b/src/uipath/eval/evaluators/llm_as_judge_evaluator.py index e4c9bffd8..f2bf2165f 100644 --- a/src/uipath/eval/evaluators/llm_as_judge_evaluator.py +++ b/src/uipath/eval/evaluators/llm_as_judge_evaluator.py @@ -51,7 +51,7 @@ class LLMJudgeMixin(BaseEvaluator[T, C, str]): """Mixin that provides common LLM judge functionality.""" system_prompt: str = LLMJudgePromptTemplates.LLM_JUDGE_SYSTEM_PROMPT - output_schema: type[BaseModel] = LLMJudgeOutputSchema + output_schema: type[BaseModel] = Field(default=LLMJudgeOutputSchema, exclude=True) actual_output_placeholder: str = "{{ActualOutput}}" expected_output_placeholder: str = "{{ExpectedOutput}}" llm_service: Callable[..., Any] | None = Field( diff --git a/src/uipath/eval/evaluators/llm_judge_output_evaluator.py b/src/uipath/eval/evaluators/llm_judge_output_evaluator.py index 400f93205..cf669f3b9 100644 --- a/src/uipath/eval/evaluators/llm_judge_output_evaluator.py +++ b/src/uipath/eval/evaluators/llm_judge_output_evaluator.py @@ -2,7 +2,7 @@ from typing import TypeVar -from pydantic import BaseModel +from pydantic import BaseModel, Field from uipath.eval.models import EvaluatorType @@ -85,7 +85,7 @@ class LLMJudgeOutputEvaluator(BaseLLMOutputEvaluator[LLMJudgeOutputEvaluatorConf """ system_prompt: str = LLMJudgePromptTemplates.LLM_JUDGE_SYSTEM_PROMPT - output_schema: type[BaseModel] = LLMJudgeOutputSchema + output_schema: type[BaseModel] = Field(default=LLMJudgeOutputSchema, exclude=True) @classmethod def get_evaluator_id(cls) -> str: @@ -105,7 +105,9 @@ class LLMJudgeStrictJSONSimilarityOutputEvaluator( system_prompt: str = ( LLMJudgePromptTemplates.LLM_JUDGE_STRICT_JSON_SIMILARITY_SYSTEM_PROMPT ) - output_schema: type[BaseModel] = LLMJudgeStrictJSONSimilarityOutputSchema + output_schema: type[BaseModel] = Field( + default=LLMJudgeStrictJSONSimilarityOutputSchema, exclude=True + ) @classmethod def get_evaluator_id(cls) -> str: diff --git a/src/uipath/eval/evaluators/llm_judge_trajectory_evaluator.py b/src/uipath/eval/evaluators/llm_judge_trajectory_evaluator.py index eac5c11b1..69b14cd1f 100644 --- a/src/uipath/eval/evaluators/llm_judge_trajectory_evaluator.py +++ b/src/uipath/eval/evaluators/llm_judge_trajectory_evaluator.py @@ -2,7 +2,7 @@ from typing import Any, TypeVar -from pydantic import BaseModel +from pydantic import BaseModel, Field from .._helpers.evaluators_helpers import trace_to_str from ..models import ( @@ -57,7 +57,9 @@ class BaseLLMTrajectoryEvaluator(LLMJudgeMixin[TrajectoryEvaluationCriteria, TC] including output extraction, prompt formatting, and evaluation criteria handling. """ - output_schema: type[BaseModel] = LLMJudgeTrajectoryOutputSchema + output_schema: type[BaseModel] = Field( + default=LLMJudgeTrajectoryOutputSchema, exclude=True + ) actual_output_placeholder: str = "{{AgentRunHistory}}" expected_output_placeholder: str = "{{ExpectedAgentBehavior}}" user_input_placeholder: str = "{{UserOrSyntheticInput}}" diff --git a/uv.lock b/uv.lock index ba0472631..dd672c3f9 100644 --- a/uv.lock +++ b/uv.lock @@ -2491,7 +2491,7 @@ wheels = [ [[package]] name = "uipath" -version = "2.7.9" +version = "2.7.10" source = { editable = "." } dependencies = [ { name = "applicationinsights" },