Skip to content

Add OmniRoute — self-hostable AI gateway with 4-tier fallback#309

Open
diegosouzapw wants to merge 1 commit intoHannibal046:mainfrom
diegosouzapw:add-omniroute
Open

Add OmniRoute — self-hostable AI gateway with 4-tier fallback#309
diegosouzapw wants to merge 1 commit intoHannibal046:mainfrom
diegosouzapw:add-omniroute

Conversation

@diegosouzapw
Copy link

Add OmniRoute

OmniRoute is a self-hostable AI gateway with:

  • 🔄 4-tier cascading fallback — automatic retry across providers
  • ⚖️ Multi-provider load balancing — distribute requests across OpenAI, Anthropic, Google, local models
  • 🔌 OpenAI-compatible API — drop-in replacement, works with any OpenAI SDK
  • 🎯 200+ models supported — GPT, Claude, Gemini, Llama, Mistral, etc.
  • 🐳 Docker-ready — single docker compose up deployment
  • 📊 Dashboard — real-time monitoring, flow visualization, provider management

GitHub: https://github.com/diegosouzapw/OmniRoute
License: MIT

Copilot AI review requested due to automatic review settings February 20, 2026 16:43
Copy link

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

This pull request adds OmniRoute to the README's LLM Inference section. OmniRoute is a self-hostable AI gateway that provides multi-provider routing, fallback mechanisms, and load balancing across 200+ AI models with an OpenAI-compatible API.

Changes:

  • Added OmniRoute entry to the LLM Inference section with a description highlighting its gateway capabilities, fallback mechanisms, and model support

💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.

- [vLLM](https://github.com/vllm-project/vllm) - A high-throughput and memory-efficient inference and serving engine for LLMs.
- [llama.cpp](https://github.com/ggerganov/llama.cpp) - LLM inference in C/C++.
- [ollama](https://github.com/ollama/ollama) - Get up and running with Llama 3, Mistral, Gemma, and other large language models.
- [OmniRoute](https://github.com/diegosouzapw/OmniRoute) - A self-hostable AI gateway with 4-tier cascading fallback, multi-provider load balancing, and OpenAI-compatible API. Supports 200+ models across OpenAI, Anthropic, Google, and local providers.
Copy link

Copilot AI Feb 20, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

OmniRoute appears to be a gateway/routing tool rather than a direct inference engine. Similar tools like "AI Gateway" (Portkey) and "TensorZero" are listed in the "LLM Applications" section (lines 510, 551), while the "LLM Inference" section primarily contains tools that perform actual inference (vLLM, llama.cpp, ollama, TGI, TensorRT-LLM). Consider moving this entry to the "LLM Applications" section, or to the "other deployment tools" subsection within "LLM Inference" (after line 461) where similar routing/serving tools like FastChat are located.

Copilot uses AI. Check for mistakes.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants