← all SaaS

LiteLLM

Open-source · self-hostable · replaces 1 SaaS tool on os-alt

BerriAI/litellm · alive · ★ 46.4k · last commit today · 2950 open issues

License: MIT

Good fit for Teams that want one OpenAI-shaped endpoint in front of many backends (mix of self-hosted + hosted Anthropic + hosted OpenAI for fallback).

Weak at Not a model server itself — you still need Ollama/vLLM/cloud APIs behind it; LiteLLM is glue, not GPU.

In a terminal? npx -y github:SolvoHQ/os-alt-cli openai-api prints the OpenAI API comparison table including LiteLLM. how the CLI works →

Replaces these SaaS

  • OpenAI API · LLM inference API

    Deploy the LiteLLM proxy with `litellm --config config.yaml`, defining model aliases like `gpt-4` → `ollama/qwen2.5:32b` and `gpt-4o-mini` → `openai/gpt-4o-mini` (real OpenAI passthrough for traffic you still want on the hosted model). Apps keep calling `gpt-4` on the OpenAI SDK with `base_url=http://litellm:4000/v1` — the routing happens in the proxy. Built-in budget caps and per-key rate limits replace OpenAI's dashboard quotas.

README badges for the SaaS this replaces

Maintainers and forks: drop a badge in your README to link readers from the SaaS-comparison page back to your repo.