Skip to content
#

prompt-routing

Here are 11 public repositories matching this topic...

Open-source LLM router & AI cost optimizer. Routes simple prompts to cheap/local models, complex ones to premium — automatically. Drop-in OpenAI-compatible proxy for Claude Code, Codex, Cursor, OpenClaw. Saves 40-70% on AI API costs. Self-hosted, no middleman.

  • Updated May 15, 2026
  • Python

Intelligent LLM router that dynamically routes prompts between local Ollama (Qwen) and cloud models (Gemini) using complexity scoring, semantic caching, and cost-aware decisioning.

  • Updated May 15, 2026
  • Python

Route prompts between local and cloud LLMs based on task complexity. Use local models (Ollama) for simple tasks, cloud APIs for complex ones. Save 80%+ on AI costs.

  • Updated May 14, 2026
  • Python

Improve this page

Add a description, image, and links to the prompt-routing topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the prompt-routing topic, visit your repo's landing page and select "manage topics."

Learn more