You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Intelligent model orchestration for Claude Code - routes queries to optimal Claude model (Haiku/Sonnet/Opus) based on complexity. It also includes many more features. If this project is working well for you and would like to support me, just help spread the word. Thanks!
Fully agentic LLM orchestrator with autonomous decision-making. This agentic system self-discovers models, learns from usage, and adapts routing strategies. Save 67% on API costs through intelligent agentic behavior. Production-ready with monitoring and self-healing.
An intelligent LLM inference gateway that dynamically routes user queries to optimal model tiers (Llama-3.1 8B/70B) based on real-time complexity, reasoning depth, and ambiguity analysis.
An intelligent, low-latency local LLM router that reduces AI costs by 30-70%. Uses a self-hosted classifier to automatically route prompts to the most cost-effective model without external API overhead.
One place to define AI tasks and map tasks to models. Stop scattering ad-hoc LLM calls across the codebase and maintain visibility, control, and cost/performance by task type.