You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Open-source LLM router & AI cost optimizer. Routes simple prompts to cheap/local models, complex ones to premium — automatically. Drop-in OpenAI-compatible proxy for Claude Code, Codex, Cursor, OpenClaw. Saves 40-70% on AI API costs. Self-hosted, no middleman.
SmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.
Stop being locked into one LLM provider. UnifyRoute is a self-hosted gateway that routes, fails over, and manages quotas across OpenAI, Anthropic, and more — with a drop-in OpenAI-compatible API.
Open-LLM Router is an OpenAI-compatible API gateway that intelligently routes LLM requests across multiple configured providers and models with features like auto-routing, failover, logging, and metrics.
LLM Router is a service that can be deployed on‑premises or in the cloud. It adds a layer between any application and the LLM provider. In real time it controls traffic, distributes a load among providers of a specific LLM, and enables analysis of outgoing requests from a security perspective (masking, anonymization, prohibited content).
An intelligent, low-latency local LLM router that reduces AI costs by 30-70%. Uses a self-hosted classifier to automatically route prompts to the most cost-effective model without external API overhead.
llm_router_services provides HTTP services that implement the core functionality used by the LLM‑Router’s plugin system. The services expose guardrail and masking capabilities through Flask applications that can be called by the corresponding plugins in llm_router_plugins.
A companion repository for llm-router containing a collection of pipeline-ready plugins. Features a masking interface for anonymizing sensitive data and a guardrail system for validating input/output safety against defined policy rules.
Successfully developed an LLM application that provides AI-powered, structured insights based on user queries. The app features a dynamic response generator with progress indicators, interactive upvote/downvote options, and a clean, engaging user interface built using Streamlit. Ideal for personalized meal, fitness, and health-related advice.
AI development environment with 90% cost savings. Routes between 8 LLM providers while defaulting to FREE local models. Production-ready with automated testing.