@@ -19,7 +19,7 @@ def test_get_threshold(self) -> None:
19
19
def test_default_threshold (self ) -> None :
20
20
thresholds = BadResponseThresholds ()
21
21
assert thresholds .get_threshold ("trustworthiness" ) == 0.7
22
- assert thresholds .get_threshold ("response_helpfulness" ) == 0.7
22
+ assert thresholds .get_threshold ("response_helpfulness" ) == 0.23
23
23
24
24
def test_unspecified_threshold (self ) -> None :
25
25
thresholds = BadResponseThresholds ()
@@ -147,7 +147,7 @@ def test_user_provided_thresholds(self, mock_project: Mock, mock_trustworthy_rag
147
147
# Test with user-provided thresholds that match evals
148
148
validator = Validator (codex_access_key = "test" , bad_response_thresholds = {"trustworthiness" : 0.6 })
149
149
assert_threshold_equal (validator , "trustworthiness" , 0.6 )
150
- assert_threshold_equal (validator , "response_helpfulness" , 0.7 )
150
+ assert_threshold_equal (validator , "response_helpfulness" , 0.23 )
151
151
152
152
# Test with extra thresholds that should raise ValueError
153
153
with pytest .raises (ValueError , match = "Found thresholds for metrics that are not available" ):
@@ -157,7 +157,7 @@ def test_default_thresholds(self, mock_project: Mock, mock_trustworthy_rag: Mock
157
157
# Test with default thresholds (bad_response_thresholds is None)
158
158
validator = Validator (codex_access_key = "test" )
159
159
assert_threshold_equal (validator , "trustworthiness" , 0.7 )
160
- assert_threshold_equal (validator , "response_helpfulness" , 0.7 )
160
+ assert_threshold_equal (validator , "response_helpfulness" , 0.23 )
161
161
162
162
def test_edge_cases (self , mock_project : Mock , mock_trustworthy_rag : Mock ) -> None : # noqa: ARG002
163
163
# Note, the `"evals"` field should not be a list of strings in practice, but an Eval from cleanlab_tlm
@@ -173,7 +173,7 @@ def test_edge_cases(self, mock_project: Mock, mock_trustworthy_rag: Mock) -> Non
173
173
# No extra Evals
174
174
validator = Validator (codex_access_key = "test" , trustworthy_rag_config = {"evals" : []})
175
175
assert_threshold_equal (validator , "trustworthiness" , 0.7 ) # Default should apply
176
- assert_threshold_equal (validator , "response_helpfulness" , 0.7 ) # Default should apply
176
+ assert_threshold_equal (validator , "response_helpfulness" , 0.23 ) # Default should apply
177
177
178
178
# Test with non-existent evals in trustworthy_rag_config
179
179
with pytest .raises (ValueError , match = "Found thresholds for metrics that are not available" ):
0 commit comments