You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Open-source LLM router & AI cost optimizer. Routes simple prompts to cheap/local models, complex ones to premium — automatically. Drop-in OpenAI-compatible proxy for Claude Code, Codex, Cursor, OpenClaw. Saves 40-70% on AI API costs. Self-hosted, no middleman.
Intelligent LLM router that dynamically routes prompts between local Ollama (Qwen) and cloud models (Gemini) using complexity scoring, semantic caching, and cost-aware decisioning.
Route prompts between local and cloud LLMs based on task complexity. Use local models (Ollama) for simple tasks, cloud APIs for complex ones. Save 80%+ on AI costs.