File tree Expand file tree Collapse file tree
Expand file tree Collapse file tree Original file line number Diff line number Diff line change 11[tool .poetry ]
22name = " deepeval"
3- version = " 2.9.6 "
3+ version = " 3.0.1 "
44description = " The LLM Evaluation Framework"
55authors = [" Jeffrey Ip <jeffreyip@confident-ai.com>" ]
66license = " Apache-2.0"
@@ -62,4 +62,4 @@ markers = [
6262 " skip_test: skip the test" ,
6363]
6464asyncio_mode = " auto"
65- asyncio_default_fixture_loop_scope = " function"
65+ asyncio_default_fixture_loop_scope = " function"
Original file line number Diff line number Diff line change 88 # Example usage of the TruthfulQA benchmark
99 # This will run the benchmark with the specified tasks and mode
1010 # using the Anthropic model.
11-
11+
1212 # Note: Ensure that you have the necessary API keys and configurations set up for the AnthropicModel.
13-
13+
1414 benchmark = TruthfulQA (
15- tasks = [TruthfulQATask .FICTION ],
16- mode = TruthfulQAMode .MC2
15+ tasks = [TruthfulQATask .FICTION ], mode = TruthfulQAMode .MC2
1716 )
1817 benchmark .evaluate (model = AnthropicModel ())
19-
You can’t perform that action at this time.
0 commit comments