OpenRouter performs much, much better than LiteLLM proxy.
In my experience, if OpenRouter offers a model, the API will be supported. They also often have inference providers available that will perform much better than the default provider. Just as an example, Z.ai is sitting at around 10 token/s for GLM 5.1 while friendly is doing 70 token/s for the same model through OpenRouter.
LiteLLM proxy also adds quite some overhead as well.
I have personally settled on a mix of Bifrost as my router which connects to OpenRouter or some other providers that I deem more privacy friendly.
LiteLLM proxy also adds quite some overhead as well.
I have personally settled on a mix of Bifrost as my router which connects to OpenRouter or some other providers that I deem more privacy friendly.