Skip to content

Commit a31ac4d

Browse files
jsondaicopybara-github
authored andcommitted
fix: fix bug where retry_timeout does not work with evaluation with experiments
PiperOrigin-RevId: 645486957
1 parent 5a6ce78 commit a31ac4d

File tree

1 file changed

+19
-3
lines changed

1 file changed

+19
-3
lines changed

vertexai/preview/evaluation/_eval_tasks.py

+19-3
Original file line numberDiff line numberDiff line change
@@ -255,6 +255,7 @@ def _evaluate_with_experiment(
255255
prompt_template: Optional[str] = None,
256256
experiment_run_name: Optional[str] = None,
257257
response_column_name: Optional[str] = None,
258+
retry_timeout: float = 600.0,
258259
) -> EvalResult:
259260
"""Runs an evaluation for the EvalTask with an experiment.
260261
@@ -270,6 +271,8 @@ def _evaluate_with_experiment(
270271
unique experiment run name is used.
271272
response_column_name: The column name of model response in the dataset. If
272273
provided, this will override the `response_column_name` of the `EvalTask`.
274+
retry_timeout: How long to keep retrying the evaluation requests for
275+
the whole evaluation dataset, in seconds.
273276
274277
Returns:
275278
The evaluation result.
@@ -285,6 +288,7 @@ def _evaluate_with_experiment(
285288
content_column_name=self.content_column_name,
286289
reference_column_name=self.reference_column_name,
287290
response_column_name=response_column_name,
291+
retry_timeout=retry_timeout,
288292
)
289293

290294
eval_result.summary_metrics = {
@@ -342,7 +346,11 @@ def evaluate(
342346
experiment=self.experiment, backing_tensorboard=False
343347
)
344348
eval_result = self._evaluate_with_experiment(
345-
model, prompt_template, experiment_run_name, response_column_name
349+
model,
350+
prompt_template,
351+
experiment_run_name,
352+
response_column_name,
353+
retry_timeout,
346354
)
347355
metadata._experiment_tracker.set_experiment(
348356
experiment=global_experiment_name, backing_tensorboard=False
@@ -352,12 +360,20 @@ def evaluate(
352360
experiment=self.experiment, backing_tensorboard=False
353361
)
354362
eval_result = self._evaluate_with_experiment(
355-
model, prompt_template, experiment_run_name, response_column_name
363+
model,
364+
prompt_template,
365+
experiment_run_name,
366+
response_column_name,
367+
retry_timeout,
356368
)
357369
metadata._experiment_tracker.reset()
358370
elif not self.experiment and global_experiment_name:
359371
eval_result = self._evaluate_with_experiment(
360-
model, prompt_template, experiment_run_name, response_column_name
372+
model,
373+
prompt_template,
374+
experiment_run_name,
375+
response_column_name,
376+
retry_timeout,
361377
)
362378
else:
363379
eval_result = _evaluation.evaluate(

0 commit comments

Comments
 (0)