Skip to content

Commit 0601b42

Browse files
author
Shriyansh Agnihotri
committed
release: version 0.0.2 🚀
1 parent a9badfb commit 0601b42

File tree

8 files changed

+249
-120
lines changed

8 files changed

+249
-120
lines changed

‎README.md‎

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -107,6 +107,7 @@ PROJECT_BASE/
107107
- `--llm-model LLM_MODEL`: Name of the LLM model to be used by the agent (recommended is `gpt-4o`, but it can take others).
108108
- `--llm-model-api-key LLM_MODEL_API_KEY`: API key for the LLM model, something like `sk-proj-k.......`.
109109

110+
110111
#### Running Hercules
111112

112113
After passing all the required parameters, the command to run Hercules should look like this:
@@ -115,6 +116,13 @@ After passing all the required parameters, the command to run Hercules should lo
115116
testzeus-hercules --input-file opt/input/test.feature --output-path opt/output --test-data-path opt/test_data --llm-model gpt-4o --llm-model-api-key sk-proj-k.......
116117
```
117118

119+
#### Supported AI Models for TestZeus-Hercules
120+
- Anthropic Haiku: Compatible with Haiku 3.5 and above.
121+
- Groq: Supports any version with function calling and coding capabilities.
122+
- Mistral: Supports any version with function calling and coding capabilities.
123+
- OpenAI: Fully compatible with GPT-4o and above. Note: OpenAI GPT-4o-mini is not supported.
124+
- Ollama: Not supported based on current testing.
125+
118126
#### Execution Flow
119127

120128
Upon running the command:

‎agents_llm_config-example.json‎

Lines changed: 53 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -4,22 +4,22 @@
44
"model_name": "mistral",
55
"model_api_key": "",
66
"model_base_url": "https://...",
7-
"system_prompt": "You are a web automation task planner....",
87
"llm_config_params": {
98
"cache_seed": null,
10-
"temperature": 0.1,
11-
"top_p": 0.1
9+
"temperature": 0.0,
10+
"top_p": 0.001,
11+
"seed":12345
1212
}
1313
},
1414
"browser_nav_agent": {
1515
"model_name": "mistral",
1616
"model_api_key": "",
1717
"model_base_url": "https://...",
18-
"system_prompt": "You will perform web navigation tasks with the functions that you have...\nOnce a task is completed, confirm completion with ##TERMINATE TASK##.",
1918
"llm_config_params": {
2019
"cache_seed": null,
21-
"temperature": 0.1,
22-
"top_p": 0.1
20+
"temperature": 0.0,
21+
"top_p": 0.001,
22+
"seed":12345
2323
}
2424
}
2525
},
@@ -47,51 +47,81 @@
4747
}
4848
}
4949
},
50-
"llama": {
50+
"ollama": {
5151
"planner_agent": {
52-
"model_name": "llama-3.1-70b-versatile",
52+
"model_name": "llama3-groq-tool-use",
53+
"model_api_key": "",
54+
"model_api_type": "ollama",
55+
"model_client_host": "http://localhost:11434",
56+
"model_native_tool_calls": true,
57+
"model_hide_tools": "if_any_run",
58+
"llm_config_params": {
59+
"cache_seed": null,
60+
"temperature": 0.0,
61+
"top_p": 0.001,
62+
"seed":12345
63+
}
64+
},
65+
"browser_nav_agent": {
66+
"model_name": "llama3-groq-tool-use",
67+
"model_api_key": "",
68+
"model_api_type": "ollama",
69+
"model_client_host": "http://localhost:11434",
70+
"model_native_tool_calls": true,
71+
"model_hide_tools": "if_any_run",
72+
"llm_config_params": {
73+
"cache_seed": null,
74+
"temperature": 0.0,
75+
"top_p": 0.001,
76+
"seed":12345
77+
}
78+
}
79+
},
80+
"groq": {
81+
"planner_agent": {
82+
"model_name": "llama3-groq-8b-8192-tool-use-preview",
5383
"model_api_key": "",
5484
"model_api_type": "groq",
55-
"system_prompt": "You are a web automation task planner....",
5685
"llm_config_params": {
5786
"cache_seed": null,
58-
"temperature": 0.1,
59-
"top_p": 0.1
87+
"temperature": 0.0,
88+
"top_p": 0.001,
89+
"seed":12345
6090
}
6191
},
6292
"browser_nav_agent": {
63-
"model_name": "llama-3.1-70b-versatile",
93+
"model_name": "llama3-groq-8b-8192-tool-use-preview",
6494
"model_api_key": "",
6595
"model_api_type": "groq",
66-
"system_prompt": "You will perform web navigation tasks with the functions that you have...\nOnce a task is completed, confirm completion with ##TERMINATE TASK##.",
6796
"llm_config_params": {
6897
"cache_seed": null,
69-
"temperature": 0.1,
70-
"top_p": 0.1
98+
"temperature": 0.0,
99+
"top_p": 0.001,
100+
"seed":12345
71101
}
72102
}
73103
},
74104
"anthropic": {
75105
"planner_agent": {
76-
"model_name": "claude-3-opus-20240229",
106+
"model_name": "claude-3-5-haiku-latest",
77107
"model_api_key": "",
78108
"model_api_type": "anthropic",
79-
"system_prompt": "You are a web automation task planner....",
80109
"llm_config_params": {
81110
"cache_seed": null,
82-
"temperature": 0.1,
83-
"top_p": 0.1
111+
"temperature": 0.0,
112+
"top_p": 0.001,
113+
"seed":12345
84114
}
85115
},
86116
"browser_nav_agent": {
87-
"model_name": "claude-3-opus-20240229",
117+
"model_name": "claude-3-5-haiku-latest",
88118
"model_api_key": "",
89119
"model_api_type": "anthropic",
90-
"system_prompt": "You will perform web navigation tasks with the functions that you have...\nOnce a task is completed, confirm completion with ##TERMINATE TASK##.",
91120
"llm_config_params": {
92121
"cache_seed": null,
93-
"temperature": 0.1,
94-
"top_p": 0.1
122+
"temperature": 0.0,
123+
"top_p": 0.001,
124+
"seed":12345
95125
}
96126
}
97127
}

0 commit comments

Comments
 (0)