@@ -19,7 +19,7 @@ def test_get_threshold(self) -> None:
1919 def test_default_threshold (self ) -> None :
2020 thresholds = BadResponseThresholds ()
2121 assert thresholds .get_threshold ("trustworthiness" ) == 0.7
22- assert thresholds .get_threshold ("response_helpfulness" ) == 0.7
22+ assert thresholds .get_threshold ("response_helpfulness" ) == 0.23
2323
2424 def test_unspecified_threshold (self ) -> None :
2525 thresholds = BadResponseThresholds ()
@@ -147,7 +147,7 @@ def test_user_provided_thresholds(self, mock_project: Mock, mock_trustworthy_rag
147147 # Test with user-provided thresholds that match evals
148148 validator = Validator (codex_access_key = "test" , bad_response_thresholds = {"trustworthiness" : 0.6 })
149149 assert_threshold_equal (validator , "trustworthiness" , 0.6 )
150- assert_threshold_equal (validator , "response_helpfulness" , 0.7 )
150+ assert_threshold_equal (validator , "response_helpfulness" , 0.23 )
151151
152152 # Test with extra thresholds that should raise ValueError
153153 with pytest .raises (ValueError , match = "Found thresholds for metrics that are not available" ):
@@ -157,7 +157,7 @@ def test_default_thresholds(self, mock_project: Mock, mock_trustworthy_rag: Mock
157157 # Test with default thresholds (bad_response_thresholds is None)
158158 validator = Validator (codex_access_key = "test" )
159159 assert_threshold_equal (validator , "trustworthiness" , 0.7 )
160- assert_threshold_equal (validator , "response_helpfulness" , 0.7 )
160+ assert_threshold_equal (validator , "response_helpfulness" , 0.23 )
161161
162162 def test_edge_cases (self , mock_project : Mock , mock_trustworthy_rag : Mock ) -> None : # noqa: ARG002
163163 # Note, the `"evals"` field should not be a list of strings in practice, but an Eval from cleanlab_tlm
@@ -173,7 +173,7 @@ def test_edge_cases(self, mock_project: Mock, mock_trustworthy_rag: Mock) -> Non
173173 # No extra Evals
174174 validator = Validator (codex_access_key = "test" , trustworthy_rag_config = {"evals" : []})
175175 assert_threshold_equal (validator , "trustworthiness" , 0.7 ) # Default should apply
176- assert_threshold_equal (validator , "response_helpfulness" , 0.7 ) # Default should apply
176+ assert_threshold_equal (validator , "response_helpfulness" , 0.23 ) # Default should apply
177177
178178 # Test with non-existent evals in trustworthy_rag_config
179179 with pytest .raises (ValueError , match = "Found thresholds for metrics that are not available" ):
0 commit comments