-
|
im trying to find examples on how to create two agents, one using OpenAi llm and another using Ollama, and pass each one its corresponding API_KEY, since most of the examples on the website define the keys (using the same name) via EXPORT Thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 4 comments 1 reply
-
|
i think i got it to work. this is an example on how to get ollama+deepseek to work |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
|
I tried this for Azure Open AI but it's not working: Also tried model: "azure/gpt-4o-mini", still doesn't work. |
Beta Was this translation helpful? Give feedback.
-
|
Agents using different LLMs is one of the most impactful optimizations for multi-agent systems — the cost and quality difference between a 7B and 70B model is enormous, and most tasks don't need the 70B. The pattern we've settled on for agent fleets: Classify before route — add a lightweight pre-call classifier (can be a regex ruleset or a small model) that maps task description → required capability tier. "Summarize this text" → fast model. "Design the architecture for X" → capable model. "Debug this subtle race condition" → reasoning model. Per-agent capability profiles — define what each specialized agent needs: Cost feedback loop — agents should know their cost per output unit (per word, per retrieval, per successful task). This creates a natural optimization pressure: agents that learn to use cheaper models for appropriate sub-tasks are more competitive in a multi-agent market. Different providers for different capabilities — Claude Haiku for retrieval and summarization (excellent cost/quality), Claude Sonnet for complex reasoning (strong mid-tier), specialized models for specific domains (coding agents → Claude or GPT-4o, math → Gemini reasoning). Provider selection and model selection are independent dimensions. We built this multi-tier routing for KinthAI's agent network: https://blog.kinthai.ai/agent-wallet-economic-models-autonomous-agents has the economic model; https://blog.kinthai.ai/openclaw-multi-tenancy-why-vm-per-user-doesnt-scale covers the infra. What's the task decomposition you're trying to optimize — is it by agent role, by subtask type, or by some other dimension? |
Beta Was this translation helpful? Give feedback.
i think i got it to work. this is an example on how to get ollama+deepseek to work