@@ -287,7 +287,7 @@ def setup_method(self):
287287 self .client = vertexai .Client (project = _TEST_PROJECT , location = _TEST_LOCATION )
288288
289289 @mock .patch .object (_evals_common , "Models" )
290- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
290+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
291291 def test_inference_with_string_model_success (
292292 self , mock_eval_dataset_loader , mock_models
293293 ):
@@ -330,7 +330,7 @@ def test_inference_with_string_model_success(
330330 assert inference_result .candidate_name == "gemini-pro"
331331 assert inference_result .gcs_source is None
332332
333- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
333+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
334334 def test_inference_with_callable_model_sets_candidate_name (
335335 self , mock_eval_dataset_loader
336336 ):
@@ -349,7 +349,7 @@ def my_model_fn(contents):
349349 assert inference_result .candidate_name == "my_model_fn"
350350 assert inference_result .gcs_source is None
351351
352- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
352+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
353353 def test_inference_with_lambda_model_candidate_name_is_none (
354354 self , mock_eval_dataset_loader
355355 ):
@@ -371,7 +371,7 @@ def test_inference_with_lambda_model_candidate_name_is_none(
371371 )
372372 assert inference_result .gcs_source is None
373373
374- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
374+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
375375 def test_inference_with_callable_model_success (self , mock_eval_dataset_loader ):
376376 mock_df = pd .DataFrame ({"prompt" : ["test prompt" ]})
377377 mock_eval_dataset_loader .return_value .load .return_value = mock_df .to_dict (
@@ -399,7 +399,7 @@ def mock_model_fn(contents):
399399 assert inference_result .gcs_source is None
400400
401401 @mock .patch .object (_evals_common , "Models" )
402- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
402+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
403403 def test_inference_with_prompt_template (
404404 self , mock_eval_dataset_loader , mock_models
405405 ):
@@ -446,7 +446,7 @@ def test_inference_with_prompt_template(
446446 assert inference_result .gcs_source is None
447447
448448 @mock .patch .object (_evals_common , "Models" )
449- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
449+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
450450 @mock .patch .object (_gcs_utils , "GcsUtils" )
451451 def test_inference_with_gcs_destination (
452452 self , mock_gcs_utils , mock_eval_dataset_loader , mock_models
@@ -500,7 +500,7 @@ def test_inference_with_gcs_destination(
500500 )
501501
502502 @mock .patch .object (_evals_common , "Models" )
503- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
503+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
504504 @mock .patch ("pandas.DataFrame.to_json" )
505505 @mock .patch ("os.makedirs" )
506506 def test_inference_with_local_destination (
@@ -552,7 +552,7 @@ def test_inference_with_local_destination(
552552 assert inference_result .gcs_source is None
553553
554554 @mock .patch .object (_evals_common , "Models" )
555- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
555+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
556556 def test_inference_from_request_column_save_to_local_dir (
557557 self , mock_eval_dataset_loader , mock_models
558558 ):
@@ -787,7 +787,7 @@ def test_inference_from_local_csv_file(self, mock_models):
787787 assert inference_result .gcs_source is None
788788
789789 @mock .patch .object (_evals_common , "Models" )
790- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
790+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
791791 def test_inference_with_row_level_config_overrides (
792792 self , mock_eval_dataset_loader , mock_models
793793 ):
@@ -972,7 +972,7 @@ def mock_generate_content_logic(*args, **kwargs):
972972 assert inference_result .gcs_source is None
973973
974974 @mock .patch .object (_evals_common , "Models" )
975- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
975+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
976976 def test_inference_with_multimodal_content (
977977 self , mock_eval_dataset_loader , mock_models
978978 ):
@@ -1048,7 +1048,7 @@ def test_inference_with_multimodal_content(
10481048 assert inference_result .candidate_name == "gemini-pro"
10491049 assert inference_result .gcs_source is None
10501050
1051- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
1051+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
10521052 @mock .patch ("vertexai._genai._evals_common.vertexai.Client" )
10531053 def test_run_inference_with_agent_engine_and_session_inputs_dict (
10541054 self ,
@@ -1136,7 +1136,7 @@ def test_run_inference_with_agent_engine_and_session_inputs_dict(
11361136 assert inference_result .candidate_name is None
11371137 assert inference_result .gcs_source is None
11381138
1139- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
1139+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
11401140 @mock .patch ("vertexai._genai._evals_common.vertexai.Client" )
11411141 def test_run_inference_with_agent_engine_and_session_inputs_literal_string (
11421142 self ,
@@ -1411,7 +1411,7 @@ def test_run_inference_with_litellm_import_error(self, mock_api_client_fixture):
14111411 @mock .patch .object (_evals_common , "_is_gemini_model" )
14121412 @mock .patch .object (_evals_common , "_is_litellm_model" )
14131413 @mock .patch .object (_evals_common , "_is_litellm_vertex_maas_model" )
1414- @mock .patch .object (_evals_metric_loaders , "EvalDatasetLoader" )
1414+ @mock .patch .object (_evals_utils , "EvalDatasetLoader" )
14151415 def test_run_inference_with_litellm_parsing (
14161416 self ,
14171417 mock_eval_dataset_loader ,
@@ -4536,9 +4536,7 @@ def test_execute_evaluation_with_openai_schema(
45364536 name = "test_metric" , prompt_template = "Evaluate: {response}"
45374537 )
45384538
4539- with mock .patch .object (
4540- _evals_metric_loaders , "EvalDatasetLoader"
4541- ) as mock_loader_class :
4539+ with mock .patch .object (_evals_utils , "EvalDatasetLoader" ) as mock_loader_class :
45424540 mock_loader_instance = mock_loader_class .return_value
45434541 mock_loader_instance .load .return_value = mock_openai_raw_data
45444542
0 commit comments