What is InferMux and how does it route inference?

InferMux routes inference requests across LLM providers. Register any backend implementing the Provider interface, and InferMux resolves models to providers automatically.

Every request is tracked: token counts, cost in USD, and a trace span reported to TokenTrace. Swap providers without changing application code.

Resources