Skip to content

Commit 2804b67

Browse files
committed
Remove OpenAI specific kwargs and try to make metric failure more obvious.
1 parent 52e49bb commit 2804b67

File tree

3 files changed

+14
-5
lines changed

3 files changed

+14
-5
lines changed

python_components/document_inference/document_inference/helpers.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -157,8 +157,7 @@ def document_inference_summary(
157157
populated_prompt,
158158
attachments=attachments,
159159
schema=DocumentSummarySchema.model_json_schema(),
160-
stream=False,
161-
reasoning_effort="minimal"
160+
stream=False
162161
)
163162
response_json = json.loads(response.text())
164163
logger.info("Inference complete. Validating response.")
@@ -183,8 +182,7 @@ def document_inference_recommendation(
183182
populated_prompt,
184183
attachments=attachments,
185184
schema=DocumentRecommendation.model_json_schema(),
186-
stream=False,
187-
reasoning_effort="minimal"
185+
stream=False
188186
)
189187
response_json = json.loads(response.text())
190188
logger.info("Inference complete. Validating response.")

python_components/evaluation/evaluation/exception/evaluation.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -143,6 +143,8 @@ async def _ceq_evaluate(self, document, exception) -> Result:
143143
input="\n\n".join(details),
144144
)
145145
metric.measure(test_case)
146+
if type(metric) is None or metric.verdicts is None:
147+
raise RuntimeError("Metric measurement failed. This is likely due to rate limiting.")
146148
details = {
147149
"verdicts": convert_model_list(metric.verdicts),
148150
"response": response,
@@ -175,6 +177,8 @@ async def _faithfulness_evaluate(self, document, exception):
175177
actual_output=[response],
176178
)
177179
metric.measure(test_case)
180+
if type(metric) is None or metric.truths is None or metric.claims is None or metric.verdicts is None:
181+
raise RuntimeError("Metric measurement failed. This is likely due to rate limiting.")
178182
details = {
179183
"truths": metric.truths,
180184
"claims": metric.claims,

python_components/evaluation/evaluation/summary/evaluation.py

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,6 @@ def evaluate(self, document: Document) -> List[Result]:
3939
"file_name": document.file_name,
4040
"inference_model": self.inference_model_name,
4141
})))
42-
4342
logger.info("Summarization complete. Performing related evaluations.")
4443
document.ai_summary = result["summary"]
4544
# Begin the DeepEval summary evaluation.
@@ -48,6 +47,14 @@ def evaluate(self, document: Document) -> List[Result]:
4847
input=document.images, actual_output=document.ai_summary
4948
)
5049
metric.measure(test_case)
50+
if (type(metric) is None
51+
or metric.truths is None
52+
or metric.claims is None
53+
or metric.assessment_questions is None
54+
or metric.coverage_verdicts is None
55+
or metric.alignment_verdicts is None
56+
):
57+
raise RuntimeError("Metric measurement failed. This is likely due to rate limiting.")
5158
details = {
5259
"truths": metric.truths,
5360
"claims": metric.claims,

0 commit comments

Comments
 (0)