@@ -426,8 +426,8 @@ def validate(
426
426
query : str ,
427
427
response : str ,
428
428
use_llm_matching : bool | NotGiven = NOT_GIVEN ,
429
- bad_response_thresholds : project_validate_params .BadResponseThresholds | NotGiven = NOT_GIVEN ,
430
429
constrain_outputs : Optional [List [str ]] | NotGiven = NOT_GIVEN ,
430
+ custom_eval_thresholds : Optional [Dict [str , float ]] | NotGiven = NOT_GIVEN ,
431
431
custom_metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
432
432
eval_scores : Optional [Dict [str , float ]] | NotGiven = NOT_GIVEN ,
433
433
options : Optional [project_validate_params .Options ] | NotGiven = NOT_GIVEN ,
@@ -451,10 +451,13 @@ def validate(
451
451
query will be recorded in the project for SMEs to answer.
452
452
453
453
Args:
454
+ custom_eval_thresholds: Optional custom thresholds for specific evals. Keys should match with the keys
455
+ in the `eval_scores` dictionary.
456
+
454
457
custom_metadata: Arbitrary metadata supplied by the user/system
455
458
456
- eval_scores: Evaluation scores to use for flagging a response as bad . If not provided, TLM
457
- will be used to generate scores.
459
+ eval_scores: Scores assessing different aspects of the RAG system . If not provided, TLM will
460
+ be used to generate scores.
458
461
459
462
options: Typed dict of advanced configuration options for the Trustworthy Language Model.
460
463
Many of these configurations are determined by the quality preset selected
@@ -575,8 +578,8 @@ def validate(
575
578
"prompt" : prompt ,
576
579
"query" : query ,
577
580
"response" : response ,
578
- "bad_response_thresholds" : bad_response_thresholds ,
579
581
"constrain_outputs" : constrain_outputs ,
582
+ "custom_eval_thresholds" : custom_eval_thresholds ,
580
583
"custom_metadata" : custom_metadata ,
581
584
"eval_scores" : eval_scores ,
582
585
"options" : options ,
@@ -967,8 +970,8 @@ async def validate(
967
970
query : str ,
968
971
response : str ,
969
972
use_llm_matching : bool | NotGiven = NOT_GIVEN ,
970
- bad_response_thresholds : project_validate_params .BadResponseThresholds | NotGiven = NOT_GIVEN ,
971
973
constrain_outputs : Optional [List [str ]] | NotGiven = NOT_GIVEN ,
974
+ custom_eval_thresholds : Optional [Dict [str , float ]] | NotGiven = NOT_GIVEN ,
972
975
custom_metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
973
976
eval_scores : Optional [Dict [str , float ]] | NotGiven = NOT_GIVEN ,
974
977
options : Optional [project_validate_params .Options ] | NotGiven = NOT_GIVEN ,
@@ -992,10 +995,13 @@ async def validate(
992
995
query will be recorded in the project for SMEs to answer.
993
996
994
997
Args:
998
+ custom_eval_thresholds: Optional custom thresholds for specific evals. Keys should match with the keys
999
+ in the `eval_scores` dictionary.
1000
+
995
1001
custom_metadata: Arbitrary metadata supplied by the user/system
996
1002
997
- eval_scores: Evaluation scores to use for flagging a response as bad . If not provided, TLM
998
- will be used to generate scores.
1003
+ eval_scores: Scores assessing different aspects of the RAG system . If not provided, TLM will
1004
+ be used to generate scores.
999
1005
1000
1006
options: Typed dict of advanced configuration options for the Trustworthy Language Model.
1001
1007
Many of these configurations are determined by the quality preset selected
@@ -1116,8 +1122,8 @@ async def validate(
1116
1122
"prompt" : prompt ,
1117
1123
"query" : query ,
1118
1124
"response" : response ,
1119
- "bad_response_thresholds" : bad_response_thresholds ,
1120
1125
"constrain_outputs" : constrain_outputs ,
1126
+ "custom_eval_thresholds" : custom_eval_thresholds ,
1121
1127
"custom_metadata" : custom_metadata ,
1122
1128
"eval_scores" : eval_scores ,
1123
1129
"options" : options ,
0 commit comments