|
50 | 50 | )
|
51 | 51 | from vertexai.preview.evaluation.metrics import pairwise_metric
|
52 | 52 | from vertexai.preview.evaluation.metrics import pointwise_metric
|
53 |
| -from vertexai.preview.evaluation.metrics import ( |
54 |
| - rubric_based_metric, |
55 |
| -) |
56 | 53 | from google.protobuf import json_format
|
57 | 54 |
|
58 | 55 |
|
@@ -260,14 +257,23 @@ def build_request(
|
260 | 257 | _default_templates.PAIRWISE_MULTIMODAL_UNDERSTANDING_RUBRIC_CRITIQUE_TEMPLATE,
|
261 | 258 | _default_templates.PAIRWISE_TEXT_QUALITY_RUBRIC_CRITIQUE_TEMPLATE,
|
262 | 259 | ):
|
263 |
| - model_based_metric_instance_input["rubrics"] = _format_rubrics( |
264 |
| - model_based_metric_instance_input["rubrics"] |
| 260 | + model_based_metric_instance_input[ |
| 261 | + constants.Dataset.RUBRICS_COLUMN |
| 262 | + ] = _format_rubrics( |
| 263 | + model_based_metric_instance_input[constants.Dataset.RUBRICS_COLUMN] |
| 264 | + ) |
| 265 | + if ( |
| 266 | + constants.Dataset.RUBRICS_COLUMN in model_based_metric_instance_input |
| 267 | + and isinstance( |
| 268 | + model_based_metric_instance_input[constants.Dataset.RUBRICS_COLUMN], |
| 269 | + List, |
| 270 | + ) |
| 271 | + ): |
| 272 | + model_based_metric_instance_input[ |
| 273 | + constants.Dataset.RUBRICS_COLUMN |
| 274 | + ] = "\n".join( |
| 275 | + model_based_metric_instance_input[constants.Dataset.RUBRICS_COLUMN] |
265 | 276 | )
|
266 |
| - if isinstance(metric, rubric_based_metric.RubricBasedMetric): |
267 |
| - if isinstance(model_based_metric_instance_input["rubrics"], List): |
268 |
| - model_based_metric_instance_input["rubrics"] = "\n".join( |
269 |
| - model_based_metric_instance_input["rubrics"] |
270 |
| - ) |
271 | 277 |
|
272 | 278 | if metric_name == constants.Metric.EXACT_MATCH:
|
273 | 279 | instance = gapic_eval_service_types.ExactMatchInput(
|
|
0 commit comments