From 1f5af0134eb4078b7b97dac5f1342542588b8056 Mon Sep 17 00:00:00 2001 From: Jaycee Li Date: Wed, 15 Apr 2026 11:51:56 -0700 Subject: [PATCH] chore: Internal cleanup PiperOrigin-RevId: 900276122 --- vertexai/_genai/evals.py | 52 ++++++++++++++++----------------- vertexai/_genai/types/common.py | 28 +++++++++--------- 2 files changed, 40 insertions(+), 40 deletions(-) diff --git a/vertexai/_genai/evals.py b/vertexai/_genai/evals.py index 7dcaacbe8a..956e48cffc 100644 --- a/vertexai/_genai/evals.py +++ b/vertexai/_genai/evals.py @@ -276,9 +276,6 @@ def _EvaluateInstancesRequestParameters_to_vertex( _EvaluationInstance_to_vertex(getv(from_object, ["instance"]), to_object), ) - if getv(from_object, ["config"]) is not None: - setv(to_object, ["config"], getv(from_object, ["config"])) - if getv(from_object, ["metric_sources"]) is not None: setv( to_object, @@ -289,6 +286,9 @@ def _EvaluateInstancesRequestParameters_to_vertex( ], ) + if getv(from_object, ["config"]) is not None: + setv(to_object, ["config"], getv(from_object, ["config"])) + return to_object @@ -464,18 +464,18 @@ def _EvaluationRunMetric_from_vertex( if getv(from_object, ["metric"]) is not None: setv(to_object, ["metric"], getv(from_object, ["metric"])) - if getv(from_object, ["metricConfig"]) is not None: + if getv(from_object, ["metricResourceName"]) is not None: setv( to_object, - ["metric_config"], - _UnifiedMetric_from_vertex(getv(from_object, ["metricConfig"]), to_object), + ["metric_resource_name"], + getv(from_object, ["metricResourceName"]), ) - if getv(from_object, ["metricResourceName"]) is not None: + if getv(from_object, ["metricConfig"]) is not None: setv( to_object, - ["metric_resource_name"], - getv(from_object, ["metricResourceName"]), + ["metric_config"], + _UnifiedMetric_from_vertex(getv(from_object, ["metricConfig"]), to_object), ) return to_object @@ -489,18 +489,18 @@ def _EvaluationRunMetric_to_vertex( if getv(from_object, ["metric"]) is not None: setv(to_object, ["metric"], getv(from_object, ["metric"])) - if getv(from_object, ["metric_config"]) is not None: + if getv(from_object, ["metric_resource_name"]) is not None: setv( to_object, - ["metricConfig"], - _UnifiedMetric_to_vertex(getv(from_object, ["metric_config"]), to_object), + ["metricResourceName"], + getv(from_object, ["metric_resource_name"]), ) - if getv(from_object, ["metric_resource_name"]) is not None: + if getv(from_object, ["metric_config"]) is not None: setv( to_object, - ["metricResourceName"], - getv(from_object, ["metric_resource_name"]), + ["metricConfig"], + _UnifiedMetric_to_vertex(getv(from_object, ["metric_config"]), to_object), ) return to_object @@ -596,9 +596,6 @@ def _GenerateInstanceRubricsRequest_to_vertex( getv(from_object, ["rubric_generation_spec"]), ) - if getv(from_object, ["config"]) is not None: - setv(to_object, ["config"], getv(from_object, ["config"])) - if getv(from_object, ["metric_resource_name"]) is not None: setv( to_object, @@ -606,6 +603,9 @@ def _GenerateInstanceRubricsRequest_to_vertex( getv(from_object, ["metric_resource_name"]), ) + if getv(from_object, ["config"]) is not None: + setv(to_object, ["config"], getv(from_object, ["config"])) + return to_object @@ -1254,8 +1254,8 @@ def _evaluate_instances( autorater_config: Optional[genai_types.AutoraterConfigOrDict] = None, metrics: Optional[list[types.MetricOrDict]] = None, instance: Optional[types.EvaluationInstanceOrDict] = None, - config: Optional[types.EvaluateInstancesConfigOrDict] = None, metric_sources: Optional[list[types.MetricSourceOrDict]] = None, + config: Optional[types.EvaluateInstancesConfigOrDict] = None, ) -> types.EvaluateInstancesResponse: """ Evaluates instances based on a given metric. @@ -1275,8 +1275,8 @@ def _evaluate_instances( autorater_config=autorater_config, metrics=metrics, instance=instance, - config=config, metric_sources=metric_sources, + config=config, ) request_url_dict: Optional[dict[str, str]] @@ -1498,8 +1498,8 @@ def _generate_rubrics( genai_types.PredefinedMetricSpecOrDict ] = None, rubric_generation_spec: Optional[genai_types.RubricGenerationSpecOrDict] = None, - config: Optional[types.RubricGenerationConfigOrDict] = None, metric_resource_name: Optional[str] = None, + config: Optional[types.RubricGenerationConfigOrDict] = None, ) -> types.GenerateInstanceRubricsResponse: """ Generates rubrics for a given prompt. @@ -1509,8 +1509,8 @@ def _generate_rubrics( contents=contents, predefined_rubric_generation_spec=predefined_rubric_generation_spec, rubric_generation_spec=rubric_generation_spec, - config=config, metric_resource_name=metric_resource_name, + config=config, ) request_url_dict: Optional[dict[str, str]] @@ -3220,8 +3220,8 @@ async def _evaluate_instances( autorater_config: Optional[genai_types.AutoraterConfigOrDict] = None, metrics: Optional[list[types.MetricOrDict]] = None, instance: Optional[types.EvaluationInstanceOrDict] = None, - config: Optional[types.EvaluateInstancesConfigOrDict] = None, metric_sources: Optional[list[types.MetricSourceOrDict]] = None, + config: Optional[types.EvaluateInstancesConfigOrDict] = None, ) -> types.EvaluateInstancesResponse: """ Evaluates instances based on a given metric. @@ -3241,8 +3241,8 @@ async def _evaluate_instances( autorater_config=autorater_config, metrics=metrics, instance=instance, - config=config, metric_sources=metric_sources, + config=config, ) request_url_dict: Optional[dict[str, str]] @@ -3470,8 +3470,8 @@ async def _generate_rubrics( genai_types.PredefinedMetricSpecOrDict ] = None, rubric_generation_spec: Optional[genai_types.RubricGenerationSpecOrDict] = None, - config: Optional[types.RubricGenerationConfigOrDict] = None, metric_resource_name: Optional[str] = None, + config: Optional[types.RubricGenerationConfigOrDict] = None, ) -> types.GenerateInstanceRubricsResponse: """ Generates rubrics for a given prompt. @@ -3481,8 +3481,8 @@ async def _generate_rubrics( contents=contents, predefined_rubric_generation_spec=predefined_rubric_generation_spec, rubric_generation_spec=rubric_generation_spec, - config=config, metric_resource_name=metric_resource_name, + config=config, ) request_url_dict: Optional[dict[str, str]] diff --git a/vertexai/_genai/types/common.py b/vertexai/_genai/types/common.py index 90278b1d7a..07d82a747f 100644 --- a/vertexai/_genai/types/common.py +++ b/vertexai/_genai/types/common.py @@ -2228,13 +2228,13 @@ class EvaluationRunMetric(_common.BaseModel): metric: Optional[str] = Field( default=None, description="""The name of the metric.""" ) - metric_config: Optional[UnifiedMetric] = Field( - default=None, description="""The unified metric used for evaluation run.""" - ) metric_resource_name: Optional[str] = Field( default=None, description="""The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}""", ) + metric_config: Optional[UnifiedMetric] = Field( + default=None, description="""The unified metric used for evaluation run.""" + ) class EvaluationRunMetricDict(TypedDict, total=False): @@ -2243,12 +2243,12 @@ class EvaluationRunMetricDict(TypedDict, total=False): metric: Optional[str] """The name of the metric.""" - metric_config: Optional[UnifiedMetricDict] - """The unified metric used for evaluation run.""" - metric_resource_name: Optional[str] """The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}""" + metric_config: Optional[UnifiedMetricDict] + """The unified metric used for evaluation run.""" + EvaluationRunMetricOrDict = Union[EvaluationRunMetric, EvaluationRunMetricDict] @@ -4412,10 +4412,10 @@ class _EvaluateInstancesRequestParameters(_common.BaseModel): instance: Optional[EvaluationInstance] = Field( default=None, description="""The instance to be evaluated.""" ) - config: Optional[EvaluateInstancesConfig] = Field(default=None, description="""""") metric_sources: Optional[list[MetricSource]] = Field( default=None, description="""The metrics used for evaluation.""" ) + config: Optional[EvaluateInstancesConfig] = Field(default=None, description="""""") class _EvaluateInstancesRequestParametersDict(TypedDict, total=False): @@ -4462,12 +4462,12 @@ class _EvaluateInstancesRequestParametersDict(TypedDict, total=False): instance: Optional[EvaluationInstanceDict] """The instance to be evaluated.""" - config: Optional[EvaluateInstancesConfigDict] - """""" - metric_sources: Optional[list[MetricSourceDict]] """The metrics used for evaluation.""" + config: Optional[EvaluateInstancesConfigDict] + """""" + _EvaluateInstancesRequestParametersOrDict = Union[ _EvaluateInstancesRequestParameters, _EvaluateInstancesRequestParametersDict @@ -5162,11 +5162,11 @@ class _GenerateInstanceRubricsRequest(_common.BaseModel): default=None, description="""Specification for how the rubrics should be generated.""", ) - config: Optional[RubricGenerationConfig] = Field(default=None, description="""""") metric_resource_name: Optional[str] = Field( default=None, description="""Registered metric resource name. If this field is set, the configuration provided in this field is used for rubric generation. The `predefined_rubric_generation_spec` and `rubric_generation_spec` fields will be ignored.""", ) + config: Optional[RubricGenerationConfig] = Field(default=None, description="""""") class _GenerateInstanceRubricsRequestDict(TypedDict, total=False): @@ -5186,12 +5186,12 @@ class _GenerateInstanceRubricsRequestDict(TypedDict, total=False): rubric_generation_spec: Optional[genai_types.RubricGenerationSpecDict] """Specification for how the rubrics should be generated.""" - config: Optional[RubricGenerationConfigDict] - """""" - metric_resource_name: Optional[str] """Registered metric resource name. If this field is set, the configuration provided in this field is used for rubric generation. The `predefined_rubric_generation_spec` and `rubric_generation_spec` fields will be ignored.""" + config: Optional[RubricGenerationConfigDict] + """""" + _GenerateInstanceRubricsRequestOrDict = Union[ _GenerateInstanceRubricsRequest, _GenerateInstanceRubricsRequestDict