InferMux routes inference requests across LLM providers. Register any backend implementing the Provider interface, and InferMux resolves models to providers automatically.
Every request is tracked: token counts, cost in USD, and a trace span reported to TokenTrace. Swap providers without changing application code.