Why Ferro Labs AI Gateway
AI gateways differ more than they look. Here's how to think about the choice.
Decision Tableโ
| Criteria | Ferro Labs OSS | LiteLLM | Portkey | Bifrost | Cloudflare AI Gateway |
|---|---|---|---|---|---|
| Language / runtime | Go | Python | Node | Go | โ |
| Latency overhead | <1 ms p99 | Varies | Cloud-dependent | <1 ms | Cloud-only |
| Self-hostable | Yes | Yes | No | Yes | No |
| OSS license | Apache 2.0 | Apache 2.0 | Proprietary | MIT | Proprietary |
| Multi-tenant | Ferro Labs Managed | Yes | Yes | No | No |
| MCP support | Yes | No | No | No | No |
| Single binary | Yes | No | No | Yes | No |
| Provider count | 29 | 100+ | Varies | Fewer | Limited |
| Active development | Yes | Yes | Yes | Yes | Yes |
Provider count is not the whole story. Ferro Labs covers the 29 providers that handle the vast majority of production traffic. LiteLLM's larger catalog includes many community-contributed integrations that vary in maintenance quality.
Which Tool Fits Your Situationโ
1. Solo dev or small team self-hostingโ
Best fit: Ferro Labs OSS
You want something that runs in minutes, has no dependencies to manage, and covers the providers you actually use. Ferro Labs AI Gateway ships as a single Go binary, works with a single config file, and supports 29 providers out of the box.
2. Python-first team already using LangChainโ
Best fit: LiteLLM
If your stack is deeply Python and you rely on tight LangChain integration, LiteLLM may fit better. It speaks native Python, integrates directly with the LangChain ecosystem, and has the largest provider catalog. The tradeoff is a heavier runtime footprint and Python process management in production.
3. Enterprise needing managed SaaSโ
Best fit: Ferro Labs Managed
When you need multi-tenant isolation, a dashboard, enterprise plugins, and someone else handling uptime, Ferro Labs Managed gives you the same gateway engine as a managed service. Portkey is also worth evaluating here if you don't need self-hosting as a fallback.
What Ferro Labs Does Betterโ
These are concrete differences, not marketing claims.
1. Zero runtime dependenciesโ
Ferro Labs AI Gateway compiles to a single Go binary. The Docker image is under 20 MB. There is no interpreter, no virtual environment, no package manager involved at deploy time. This matters when you are running in constrained environments or want reproducible deploys.
2. Eight routing strategiesโ
Ferro Labs ships with 8 routing strategies including content-based routing and A/B testing. You can route by model, cost, latency, or the content of the request itself โ without writing custom middleware.
3. Agentic MCP loop built-inโ
The gateway runs a full MCP tool loop internally. Your client sends a normal chat completion request; the gateway handles tool discovery, execution, and result injection. No client-side changes required.
The built-in MCP loop means you can add tool-use capabilities to any application that speaks the OpenAI API format, even if that application has no concept of function calling.
4. Single config file, 29 providersโ
One configuration file activates any combination of 29 providers. Ferro Labs AI Gateway discovers credentials from environment variables automatically โ no per-provider setup code.
5. Apache 2.0 with no surprisesโ
Ferro Labs OSS is Apache 2.0 licensed. There are no "community" vs. "enterprise" edition splits in the open-source core, and no license changes to worry about as you scale.
Bifrost is MIT-licensed, which is also permissive. LiteLLM is Apache 2.0 as well. Portkey and Cloudflare AI Gateway are proprietary. Pick the license model that matches your organization's requirements.