LiteLLM
Open-source · self-hostable · replaces 1 SaaS tool on os-alt
BerriAI/litellm · alive · ★ 46.4k · last commit today · 2950 open issues
License: MIT
Good fit for Teams that want one OpenAI-shaped endpoint in front of many backends (mix of self-hosted + hosted Anthropic + hosted OpenAI for fallback).
Weak at Not a model server itself — you still need Ollama/vLLM/cloud APIs behind it; LiteLLM is glue, not GPU.
In a terminal? npx -y github:SolvoHQ/os-alt-cli openai-api prints
the OpenAI API comparison table including LiteLLM.
how the CLI works →
Replaces these SaaS
- OpenAI API · LLM inference API
Deploy the LiteLLM proxy with `litellm --config config.yaml`, defining model aliases like `gpt-4` → `ollama/qwen2.5:32b` and `gpt-4o-mini` → `openai/gpt-4o-mini` (real OpenAI passthrough for traffic you still want on the hosted model). Apps keep calling `gpt-4` on the OpenAI SDK with `base_url=http://litellm:4000/v1` — the routing happens in the proxy. Built-in budget caps and per-key rate limits replace OpenAI's dashboard quotas.
README badges for the SaaS this replaces
Maintainers and forks: drop a badge in your README to link readers from the SaaS-comparison page back to your repo.